Abstract
Both intensional and extensional background knowledge have previously been used in inductive problems to complement the training set used for a task. In this research, we propose to explore the usefulness, for inductive learning, of a new kind of intensional background knowledge: the inter-relationships or conditional probability distributions between subsets of attributes. Such information could be mined from publicly available knowledge sources but including only some of the attributes involved in the inductive task at hand. The purpose of our work is to show how this information can be useful in inductive tasks, and under what circumstances. We will consider injection of background knowledge into Bayesian Networks and explore its effectiveness on training sets of different sizes. We show that this additional knowledge not only improves the estimate of classification accuracy, it also reduces the variance in the accuracy of the model.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Blake, C., Merz, C.: UCI repository of machine learning databases. Univ. of California, Irvine, http://www.ics.uci.edu/~mlearn/MLRepository.html
Cheng, J.: BN powerpredictor, http://webdocs.cs.ualberta.ca/~jcheng/bnsoft.htm
Cheng, J., Greiner, R.: Learning bayesian belief network classifiers: Algorithms and system. In: Stroulia, E., Matwin, S. (eds.) Canadian AI 2001. LNCS (LNAI), vol. 2056, pp. 141–151. Springer, Heidelberg (2001)
Clark, P., Matwin, S.: Learning domain theories using abstract beckground knowledge. In: Brazdil, P.B. (ed.) ECML 1993. LNCS, vol. 667, pp. 360–365. Springer, Heidelberg (1993)
Flach, P.A.: From extensional to intensional knowledge: Inductive logic programming techniques and their application to deductive databases. In: Kifer, M., Voronkov, A., Freitag, B., Decker, H. (eds.) Dagstuhl Seminar 1997, DYNAMICS 1997, and ILPS-WS 1997. LNCS, vol. 1472, pp. 356–387. Springer, Heidelberg (1998)
Heckerman, D.: A tutorial on learning with bayesian networks. Technical Report MSR-TR-95-06, Microsoft Research (1995)
Mitchell, T.M.: Machine learning. McGraw-Hill, New York (1997)
Seifi, F., Drummond, C., Japkowicz, N., Matwin, S.: Improving Bayesian Learning Using Public Knowledge, p. 12, http://www.archive.org/details/ImprovingBayesianLearningUsingPublicKnowledge
Wu, P., Dietterich, T.G.: Improving svm accuracy by training on auxiliary data sources. In: Brodley, C.E. (ed.) ICML. ACM International Conference Proceeding Series, vol. 69. ACM, New York (2004)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Seifi, F., Drummond, C., Japkowicz, N., Matwin, S. (2010). Improving Bayesian Learning Using Public Knowledge. In: Farzindar, A., Kešelj, V. (eds) Advances in Artificial Intelligence. Canadian AI 2010. Lecture Notes in Computer Science(), vol 6085. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-13059-5_44
Download citation
DOI: https://doi.org/10.1007/978-3-642-13059-5_44
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-13058-8
Online ISBN: 978-3-642-13059-5
eBook Packages: Computer ScienceComputer Science (R0)