Abstract
Generic ensemble methods can achieve excellent learning performance, but are not good candidates for active learning because of their different design purposes. We investigate how to use diversity of the member classifiers of an ensemble for efficient active learning. We empirically show, using benchmark data sets, that (1) to achieve a good (stable) ensemble, the number of classifiers needed in the ensemble varies for different data sets; (2) feature selection can be applied for classifier selection from ensembles to construct compact ensembles with high performance. Benchmark data sets and a real-world application are used to demonstrate the effectiveness of the proposed approach.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Blake, C.L., Merz, C.J.: UCI repository of machine learning databases (1998), http://www.ics.uci.edu/~mlearn/MLRepository.html
Blum, A., Mitchell, T.: Combining labeled and unlabeled data with co-training. In: COLT: Proc. of Workshop on Computational Learning Theory, Morgan Kaufmann Publishers, San Francisco (1998)
Breiman, L.: Bagging predictors. Machine Learning 24, 123–140 (1996)
Dietterich, T.G.: Ensemble methods in machine learning. In: Proc. of Intl. Workshop on Multiple Classifier Systems, pp. 1–15. Springer, Heidelberg (2000)
Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: Proc. of Intl. Conf. on Machine Learning, pp. 148–156. Morgan Kaufmann, San Francisco (1996)
Freund, Y., Seung, H., Shamir, E., Tishby, N.: Selective sampling using the query by committee algorithm. Machine Learning 28, 133–168 (1997)
Lewis, D., Gale, W.: A sequential algorithm for training text classifiers. In: Proc. of ACM-SIGR Conference on Research and Development in Information Retrieval, pp. 3–12 (1994)
Mandvikar, A., Liu, H.: Class-Specific Ensembles for Active Learning in Digital Imagery. In: Proc. of SIAM Intl. Conf. on Data Mining (2004) (in press)
McCallum, A., Nigam, K.: Employing EM in pool-based active learning for text classification. In: Proc. of Intl. Conf. on Machine Learning, pp. 350–358 (1998)
Nigam, K., Mccallum, A.K., Thrun, S., Mitchell, T.: Text classification from labeled and unlabeled documents usingEM. Machine Learning 39, 103–134 (2000)
Roy, N., McCallum, A.: Toward optimal active learning through sampling estimation of error reduction. In: Proc. of Intl. Conf. On Machine Learning (2001)
Seung, H.S., Opper, M., Sompolinsky, H.: Query by committee. In: Proc. of Workshop on Computational Learning Theory, pp. 287–294. ACM Press, New York (1992)
Witten, I.H., Frank, E.: Data Mining – Practical Machine Learning Tools and Techniques with JAVA Implementations. Morgan Kaufmann Publishers, San Francisco (2000)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Mandvikar, A., Liu, H., Motoda, H. (2004). Compact Dual Ensembles for Active Learning. In: Dai, H., Srikant, R., Zhang, C. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2004. Lecture Notes in Computer Science(), vol 3056. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-24775-3_37
Download citation
DOI: https://doi.org/10.1007/978-3-540-24775-3_37
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-22064-0
Online ISBN: 978-3-540-24775-3
eBook Packages: Springer Book Archive