Abstract
This paper describes a method for selecting training examples for a partial memory learning system. The method selects extreme examples that lie at the boundaries of concept descriptions and uses these examples with new training examples to induce new concept descriptions. Forgetting mechanisms also may be active to remove examples from partial memory that are irrelevant or outdated for the learning task. Using an implementation of the method, we conducted a lesion study and a direct comparison to examine the effects of partial memory learning on predictive accuracy and on the number of training examples maintained during learning. These experiments involved the STAGGER Concepts, a synthetic problem, and two real-world problems: a blasting cap detection problem and a computer intrusion detection problem. Experimental results suggest that the partial memory learner notably reduced memory requirements at the slight expense of predictive accuracy, and tracked concept drift as well as other learners designed for this task.
Article PDF
Similar content being viewed by others
References
Aha, D., Kibler, D., & Albert, M. (1991). Instance-based learning algorithms. Machine Learning, 6, 37–66.
Baim, P. (1988). A method for attribute selection in inductive learning systems. IEEE Transactions on Pattern Analysis and Machine Intelligence, 10(6), 888–896.
Blake, C., Keogh, E., & Merz, C. (1998). UCI repository of machine learning databases [On-line]. Department of Information and Computer Sciences, University of California, Irvine. Available: http://www.ics.uci.edu/~mlearn/mlrepository.html
Bloedorn, E. & Michalski, R. (1998). Data-driven constructive induction. IEEE Intelligent Systems, 13(2), 30–37.
Bloedorn, E., Wnek, J., Michalski, R., & Kaufman, K. (1993). AQ-17-A multistrategy learning system: The method and user's guide. Reports of the Machine Learning and Inference Laboratory MLI 93–12. Fairfax, VA: George Mason University.
Clark, P. & Niblett, T. (1989). The CN2 induction algorithm. Machine Learning, 3, 261–284.
Davis, J. (1981). CONVART: A program for constructive induction on time dependent data. Master's Thesis, Department of Computer Science, University of Illinois, Urbana-Champaign.
Elio, R. & Watanabe, L. (1991). An incremental deductive strategy for controlling constructive induction in learning from examples. Machine Learning, 7, 7–44.
Fisher, R. (1936). The use of multiple measurements in taxonomic problems. Annals of Eugenics, 7, 179–188.
Frisch, A. (1995). Essential system administration (2nd ed.). Sebastopol, CA: O'Reilly & Associates.
Iba, W., Woogulis, J., & Langley, P. (1988). Trading simplicity and coverage in incremental concept learning. In Proceedings of the Fifth International Conference on Machine Learning (pp. 73–79). San Francisco, CA: Morgan Kaufmann.
Kaufman, K. & Michalski, R. (1998). AQ-18: An environment for learning and natural induction. Reports of the Machine Learning and Inference Laboratory MLI 98–12. Fairfax, VA: George Mason University.
Kerber, R. (1992). ChiMerge: discretization of numeric attributes. In Proceedings of the Tenth National Conference on Artificial Intelligence (pp. 123–128). Menlo Park, CA: AAAI Press.
Kibler, D. & Langley, P. (1990). Machine learning as an experimental science. In Dietterich, T. & Shavlik, J. (Eds.), Readings in Machine Learning. San Francisco, CA: Morgan Kaufmann.
Krizakova, I. & Kubat, M. (1992). FAVORIT: Concept formation with ageing of knowledge. Pattern Recognition Letters, 13, 19–25.
Kubat, M. & Krizakova, I. (1992). Forgetting and aging of knowledge in concept formation. Applied Artificial Intelligence, 6, 195–206.
Lebowitz, M. (1987). Experiments with incremental concept formation: UNIMEM. Machine Learning, 2, 103–138.
Littlestone, N. (1991). Redundant noisy attributes, attribute errors, and linear-threshold learning using Winnow. In Proceedings of the Fourth Annual Workshop on Computational Learning Theory (pp. 147–156). San Francisco, CA: Morgan Kaufmann.
Maloof, M. (1996). Progressive partial memory learning. Doctoral Dissertation, School of Information Technology and Engineering, George Mason University, Fairfax, VA.
Maloof, M., Duric, Z., Michalski, R., & Rosenfeld, A. (1996). Recognizing blasting caps in X-ray images. In Proceedings of the Image Understanding Workshop (pp. 1257–1261). San Francisco, CA: Morgan Kaufmann.
Maloof, M. & Michalski, R. (1995). A method for partial-memory incremental learning and its application to computer intrusion detection. In Proceedings of the Seventh IEEE International Conference on Tools with Artificial Intelligence (pp. 392–397). Los Alamitos, CA: IEEE Press.
Maloof, M. & Michalski, R. (1997). Learning symbolic descriptions of shape for object recognition in X-ray images. Expert Systems with Applications, 12(1), 11–20.
Michalski, R. (1969). On the quasi-minimal solution of the general covering problem. In Proceedings of the Fifth International Symposium on Information Processing, Vol. A3 (pp. 125–128).
Michalski, R. (1980). Pattern recognition as rule-guided inductive inference. IEEE Transactions onPattern Analysis and Machine Intelligence, 2(4), 349–361.
Michalski, R. (1983). A theory and methodology of inductive learning. In Michalski, R., Carbonell, J., & Mitchell, T. (Eds.), Machine Learning: An Artificial Intelligence Approach,Vol. 1. San Francisco, CA: Morgan Kaufmann.
Michalski, R. & Larson, J. (1983). Incremental generation of VL1 hypotheses: The underlying methodology and the description of program AQ-11. Technical report UIUCDCS-F–83–905, Department of Computer Science, University of Illinois, Urbana-Champaign.
Quinlan, J. (1993). C4.5: Programs for Machine Learning. San Francisco, CA: Morgan Kaufmann.
Reinke, R. & Michalski, R. (1988). Incremental learning of concept descriptions: A method and experimental results. In Hayes, J., Michie, D., & Richards, J. (Eds.), Machine Intelligence 11. Oxford: Clarendon Press.
Salganicoff, M. (1993). Density-adaptive learning and forgetting. In Proceedings of the Tenth International Conference on Machine Learning (pp. 276–283). San Francisco, CA: Morgan Kaufmann.
Schlimmer, J. & Granger, R. (1986). Beyond incremental processing: Tracking concept drift. In Proceedings of the Fifth National Conference on Artificial Intelligence (pp. 502–507). Menlo Park, CA: AAAI Press.
Utgoff, P. (1988). ID5: An incremental ID3. In Proceedings of the Fifth International Conference on Machine Learning (pp. 107–120). San Francisco, CA: Morgan Kaufmann.
Watanabe, L. & Elio, R. (1987). Guiding constructive induction for incremental learning from examples. In Proceedings of the Tenth International Joint Conference on Artificial Intelligence (pp. 293–296). San Francisco, CA: Morgan Kaufmann.
Widmer, G. (1997). Tracking context changes through meta-learning. Machine Learning, 27, 259–286.
Widmer, G.& Kubat, M. (1996). Learning in the presence of concept drift and hidden contexts. Machine Learning, 23, 69–101.
Wnek, J., Kaufman, K., Bloedorn, E., & Michalski, R. (1995). Selective induction learning system AQ-15c: The method and user's guide. Reports of the Machine Learning and Inference Laboratory MLI 95–4. Fairfax, VA
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Maloof, M.A., Michalski, R.S. Selecting Examples for Partial Memory Learning. Machine Learning 41, 27–52 (2000). https://doi.org/10.1023/A:1007661119649
Issue Date:
DOI: https://doi.org/10.1023/A:1007661119649