Skip to main content
Log in

Improving bagging performance through multi-algorithm ensembles

  • Research Article
  • Published:
Frontiers of Computer Science Aims and scope Submit manuscript

Abstract

Working as an ensemble method that establishes a committee of classifiers first and then aggregates their outcomes through majority voting, bagging has attracted considerable research interest and been applied in various application domains. It has demonstrated several advantages, but in its present form, bagging has been found to be less accurate than some other ensemble methods. To unlock its power and expand its user base, we propose an approach that improves bagging through the use of multi-algorithm ensembles. In a multi-algorithm ensemble, multiple classification algorithms are employed. Starting from a study of the nature of diversity, we show that compared to using different training sets alone, using heterogeneous algorithms together with different training sets increases diversity in ensembles, and hence we provide a fundamental explanation for research utilizing heterogeneous algorithms. In addition, we partially address the problem of the relationship between diversity and accuracy by providing a non-linear function that describes the relationship between diversity and correlation. Furthermore, after realizing that the bootstrap procedure is the exclusive source of diversity in bagging, we use heterogeneity as another source of diversity and propose an approach utilizing heterogeneous algorithms in bagging. For evaluation, we consider several benchmark data sets from various application domains. The results indicate that, in terms of F1-measure, our approach outperforms most of the other state-of-the-art ensemble methods considered in experiments and, in terms of mean margin, our approach is superior to all the others considered in experiments.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Similar content being viewed by others

References

  1. Rokach L. Pattern Classification Using Ensemble Methods. Hackensack: World Scientific Pub. Co. Inc, 2010

    MATH  Google Scholar 

  2. Breiman L. Bagging predictors. Machine learning, 1996, 24(2): 123–140

    MathSciNet  MATH  Google Scholar 

  3. Pinheiro C A R, Evsukoff A, Ebecken N F F. Revenue recovering with insolvency prevention on a Brazilian telecom operator. ACM SIGKDD Explorations Newsletter, 2006, 8(1): 65–70

    Article  Google Scholar 

  4. Lemmens A, Croux C. Bagging and boosting classification trees to predict churn. Journal of Marketing Research, 2006, 43(2): 276–286

    Article  Google Scholar 

  5. Perlich C, Rosset S, Lawrence R D, Zadrozny B. High-quantile modeling for customer wallet estimation and other applications. In: Proceedings of the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2007, 977–985

  6. Lasota T, Telec Z, Trawiński B, Trawiński K. A multi-agent system to assist with real estate appraisals using bagging ensembles. In: Proceedings of the 1st International Conference on Computational Collective Intelligence — Semantic Web, Social Networks and Multiagent Systems. 2009, 813–824

  7. Paleologo G, Elisseeff A, Antonini G. Subagging for credit scoring models. European Journal of Operational Research, 2010, 201(2): 490–499

    Article  Google Scholar 

  8. Hothorn T, Lausen B. Bagging tree classifiers for laser scanning images: a data- and simulation-based strategy. Artificial Intelligence in Medicine, 2003, 27(1): 65–79

    Article  Google Scholar 

  9. Nunkesser R, Bernholt T, Schwender H, Ickstadt K, Wegener I. Detecting high-order interactions of single nucleotide polymorphisms using genetic programming. Bioinformatics, 2007, 23(24): 3280–3288

    Article  Google Scholar 

  10. Lu C, Devos A, Suykens J A K, Arus C, Van Huffel S. Bagging linear sparse bayesian learning models for variable selection in cancer diagnosis. IEEE Transactions on Information Technology in Biomedicine, 2007, 11(3): 338–347

    Article  Google Scholar 

  11. Larios N, Deng H, Zhang W, Sarpola M, Yuen J, Paasch R, Moldenke A, Lytle D A, Ruiz-Correa S, Mortensen E N, Shapiro L G, Dietterich T G. Automated insect identification through concatenated histograms of local appearance features: feature vector generation and region detection for deformable objects. Machine Vision and Applications, 2008, 19(2): 105–123

    Article  Google Scholar 

  12. Stepinski T F, Ghosh S, Vilalta R. Machine learning for automatic mapping of planetary surfaces. In: Proceedings of the 19th National Conference on Innovative Applications of Artificial Intelligence. 2007, 1807–1812

  13. Wu F, Weld D S. Autonomously semantifying wikipedia. In: Proceedings of the 16th ACM Conference on Information and Knowledge Management. 2007, 41–50

  14. Singh K, İpek E, McKee S A, de Supinski B R, Schulz M, Caruana R. Predicting parallel application performance via machine learning approaches. Concurrency and Computation: Practice and Experience, 2007, 19(17): 2219–2235

    Article  Google Scholar 

  15. Braga P L, Oliveira A L I, Ribeiro G H T, Meira S R L. Bagging predictors for estimation of software project effort. In: Proceedings of International Joint Conference on Neural Networks. 2007, 1595–1600

  16. Aljamaan H I, Elish M O. An empirical study of bagging and boosting ensembles for identifying faulty classes in object-oriented software. In: Proceedings of IEEE Symposium on Computational Intelligence and Data Mining. 2009, 187–194

  17. Hulth A. Improved automatic keyword extraction given more linguistic knowledge. In: Proceedings of the 2003 Conference on Empirical Methods in Natural Language Processing. 2003, 216–223

  18. Kurogi S, Nedachi N, Funatsu Y. Reproduction and recognition of vowel signals using single and bagging competitive associative nets. In: Proceedings of the 14th International Conference on Neural Information Processing, Part II. 2008, 40–49

  19. Kurogi S, Sato S, Ichimaru K. Speaker recognition using pole distribution of speech signals obtained by bagging CAN2. In: Proceedings of the 16th International Conference on Neural Information Processing, Part I. 2009, 622–629

  20. Boinee P, De Angelis A, Foresti G L. Ensembling classifiers — an application to image data classification from Cherenkov telescope experiment. In: Proceedings of International Enformatika Conference. 2005, 394–398

  21. Wang Y, Wang Y, Jain A K, Tan T. Face verification based on bagging RBF networks. In: Proceedings of International Conference on Advances in Biometrics. 2006, 69–77

  22. Quinlan J R. Bagging, boosting, and c4.5. In: Proceedings of the 13th National Conference on Artificial Intelligence. 1996, 725–730

  23. Maclin R, Opitz D. An empirical evaluation of bagging and boosting. In: Proceedings of the 14th National Conference on Artificial Intelligence and the 9th Conference on Innovative Applications of Artificial Intelligence. 1997, 546–551

  24. Opitz D W, Maclin R. Popular ensemble methods: an empirical study. Journal of Artificial Intelligence Research, 1999, 11: 169–198

    MATH  Google Scholar 

  25. Kotsiantis S B, Pintelas P E. Combining bagging and boosting. International Journal of Computational Intelligence, 2004, 1(4): 324–333

    Google Scholar 

  26. Banfield R E, Hall L O, Bowyer K W, Kegelmeyer WP. A comparison of decision tree ensemble creation techniques. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2007, 29(1): 173–180

    Article  Google Scholar 

  27. Tumer K, Ghosh J. Error correlation and error reduction in ensemble classifiers. Connection Science, 1996, 8(3): 385–404

    Article  Google Scholar 

  28. Breiman L. Random forests. Machine learning, 2001, 45(1): 5–32

    Article  MATH  Google Scholar 

  29. Tan P N, Steinbach M, Kumar V. Introduction to Data Mining. Boston: Addison Wesley, 2005

    Google Scholar 

  30. Ting K M, Wells J R, Tan S C, Teng S W, Webb G I. Featuresubspace aggregating: ensembles for stable and unstable learners. Machine Learning, 2011, 82(3): 375–397

    Article  Google Scholar 

  31. Wang Q, Zhang L. Ensemble learning based on multi-task class labels. In: Proceedings of the 14th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining. 2010, 464–475

  32. Tikk D, Kardkovács Z T, Szidarovszky F P. Voting with a parameterized veto strategy: solving the KDD cup 2006 problem by means of a classifier committee. ACM SIGKDD Explorations Newsletter, 2006, 8(2): 53–62

    Article  Google Scholar 

  33. Lo H Y, Chang KW, Chen S T, Chiang T H, Ferng C S, Hsieh C J, Ko Y K, Kuo T T, Lai H C, Lin K Y, Wang C H, Yu H F, Lin C J, Lin H T, Lin S D. An ensemble of three classifiers for KDD cup 2009: expanded linear model, heterogeneous boosting, and selective naive Bayes. Journal of Machine Learning Research — Proceedings Track, 2009, 7: 57–64

    Google Scholar 

  34. Niculescu-Mizil A, Perlich C, Swirszcz G, Sindhwani V, Liu Y, Melville P, Wang D, Xiao J, Hu J, Singh M, Shang WX, Zhu Y F. Winning the KDD cup orange challenge with ensemble selection. Journal of Machine Learning Research — Proceedings Track, 2009, 7: 23–34

    Google Scholar 

  35. Xie J, Rojkova V, Pal S, Coggeshall S. A combination of boosting and bagging for KDD cup 2009—fast scoring on a large database. Journal of Machine Learning Research — Proceedings Track, 2009, 7: 35–43

    Google Scholar 

  36. Yu H F, Lo H Y, Hsieh H P, Lou J K, McKenzie T G, Chou JW, Chung P H, Ho C H, Chang C F, Wei Y H, Weng J Y, Yan E S, Chang C W, Kuo T T, Lo YC, Chang P T, Po C, Wang C Y, Huang Y H, Hung CW, Ruan YX, Lin YS, Lin SD, Lin HT, Lin C J. Feature engineering and classifier ensemble for KDD cup 2010. Journal of Machine Learning Research: Workshop and Conference Proceedings, 2010, 1: 1–16

    Google Scholar 

  37. Hsu KW, Srivastava J. Improving bagging performance through multialgorithm ensembles. In: Proceedings of the 1st Doctoral Symposium on Data Mining, in conjunction with the 15th Pacific-Asia Conference on Knowledge Discovery and Data Mining. 2011

  38. Hsu K W, Srivastava J. Diversity in combinations of heterogeneous classifiers. In: Proceedings of the 13th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining. 2009, 923–932

  39. Hsu K W, Srivastava J. An empirical study of applying ensembles of heterogeneous classifiers on imperfect data. In: Proceedings of PAKDD 2009 International Workshops on New Frontiers in Applied Data Mining. 2010, 28–39

  40. Hsu K W, Srivastava J. Relationship between diversity and correlation in multi-classifier systems. In: Proceedings of the 14th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining. 2010, 500–506

  41. Hsu K W. Applying bagging with heterogeneous algorithms to health care data. In: Proceedings of the 1st Workshop on Data Mining for Healthcare Management, in conjunction with the 14th Pacific-Asia Conference on Knowledge Discovery and Data Mining. 2010

  42. Meynet J. Information theoretic combination of classifiers with application to face detection. Dissertation for the Doctoral Degree. Lausanne: EPFL, 2007

    Google Scholar 

  43. Kuncheva L I, Whitaker C J. Ten measures of diversity in classifier ensembles: limits for two classifiers. In: Proceedings of IEEEWorkshop on Intelligent Sensor Processing. 2001, 1–10

  44. Aksela M. Comparison of classifier selection methods for improving committee performance. In: Proceedings of the 4th International Conference on Multiple Classifier Systems. 2003, 84–93

  45. Banfield R E, Hall L O, Bowyer K W, Kegelmeyer WP. A new ensemble diversity measure applied to thinning ensembles. In: Proceedings of the 4th International Conference on Multiple Classifier Systems. 2003, 306–316

  46. Kuncheva L I, Whitaker C J. Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Machine Learning, 2003, 51(2): 181–207

    Article  MATH  Google Scholar 

  47. Kuncheva L I. That elusive diversity in classifier ensembles. In: Proceedings of the 1st Iberian Conference on Pattern Recognition and Image Analysis. 2003, 1126–1138

  48. Kuncheva L I. Combining Pattern Classifiers: Methods and Algorithms. New York: Wiley-Interscience, 2004

    Book  MATH  Google Scholar 

  49. Brown G, Wyatt J L, Harris R, Yao X. Diversity creation methods: a survey and categorisation. Information Fusion, 2005, 6(1): 5–20

    Article  Google Scholar 

  50. Liu W, Wu Z, Pan G. An entropy based diversity measure for classifier combining and its application to face classifier ensemble thinning. In: Proceedings of the 5th Chinese Conference on Biometric Recognition. 2004, 118–124

  51. Aksela M, Laaksonen J. Using diversity of errors for selecting members of a committee classifier. Pattern Recognition, 2006, 39(4): 608–623

    Article  MATH  Google Scholar 

  52. Fan T G, Zhu Y, Chen J M. A new measure of classifier diversity in multiple classifier system. In: Proceedings of International Conference on Machine Learning and Cybernetics. 2008, 18–21

  53. Ghosh J. Multiclassifier systems: back to the future. In: Proceedings of the 3rd International Workshop on Multiple Classifier Systems. 2003, 1–15

  54. Brown G. Ensemble learning. Encyclopedia of Machine Learning. Heidelberg: Springer Press, 2010

    Google Scholar 

  55. Bühlmann P, Yu B. Analyzing bagging. Annals of Statistics, 2003, 30(4): 927–961

    Google Scholar 

  56. Davison A C, Hinkley D V. Bootstrap Methods and Their Application. Cambridge: Cambridge University Press, 1997

    MATH  Google Scholar 

  57. Frank A, Asuncion A. UCI machine learning repository. 2010, http://archive.ics.uci.edu/ml

  58. Wu X, Kumar V, Quinlan J R, Ghosh J, Yang Q, Motoda H, McLachlan G J, Ng A, Liu B, Yu P S, Zhou Z H, Steinbach M, Hand D J, Steinberg D. Top 10 algorithms in data mining. Knowledge and Information Systems, 2008, 14(1): 1–37

    Article  Google Scholar 

  59. Platt J C. Fast training of support vector machines using sequential minimal optimization. In: Schölkopf B, Burges C J C, Smola A J, eds. Advances in Kernel Methods — Support Vector Learning. Cambridge: MIT Press, 1998

    Google Scholar 

  60. Schapire R E. The strength of weak learnability. Machine Learning, 1990, 5(2): 197–227

    Google Scholar 

  61. Melville P, Mooney R J. Constructing diverse classifier ensembles using artificial training examples. In: Proceedings of the 18th International Joint Conference on Artificial Intelligence. 2003, 505–510

  62. Melville P, Mooney R J. Creating diversity in ensembles using artificial data. Information Fusion, 2004, 6(1): 99–111

    Article  Google Scholar 

  63. Wolpert D H. Stacked generalization. Neural Networks, 1992, 5(2): 241–259

    Article  MathSciNet  Google Scholar 

  64. Seewald A K. How to make stacking better and faster while also taking care of an unknown weakness. In: Proceedings of the 19th International Conference on Machine Learning. 2003, 554–561

  65. Rodriguez J J, Kuncheva L I, Alonso C J. Rotation forest: a new classifier ensemble method. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2006, 28(10): 1619–1630

    Article  Google Scholar 

  66. Kuncheva L I, Rodriguez J J. An experimental study on rotation forest ensembles. In: Proceedings of the 7th International Conference on Multiple Classifier Systems. 2007, 459–468

  67. Ho T K. The random subspace method for constructing decision forests. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1998, 20(8): 832–844

    Article  Google Scholar 

  68. Quinlan J R. Learning with continuous classes. In: Proceedings of the 5th Australian Joint Conference on Artificial Intelligence. 1992, 343–348

  69. Quinlan J R. C4.5: Programs for Machine Learning. San Francisco: Morgan Kaufmann, 1993

    Google Scholar 

  70. Ting K M, Witten I H. Stacking bagged and dagged models. In: Proceedings of the 14th International Conference on Machine Learning. 1997, 367–375

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kuo-Wei Hsu.

Additional information

Kuo-Wei (David) Hsu is currently an assistant professor in the Department of Computer Science at the Chengchi University, China. He earned his PhD from the University of Minnesota, USA. Before he entered the PhD program, he worked as an information engineer in the Taiwan University Hospital, China. Prior to that, he obtained his MS and BS from Taiwan University, and Chung Hsing University, China, respectively. His current research interests include data mining, database systems, and software engineering.

Jaideep Srivastava is a professor of Computer Science and Engineering at the University of Minnesota, USA. He has established and led a laboratory that conducts research in databases, multimedia systems, and data mining. Dr. Srivastava has an active collaboration with the technology industry, both for research and technology transfer, and is an often-invited participant in technical and technology strategy forums. The US federal government has solicited his opinion on computer science research as an expert witness. Dr. Srivastava has a BTech. degree from the Indian Institute of Technology, Kanpur, India, and MS and PhD from the University of California, Berkeley. He has been elected as a Fellow of the IEEE.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Hsu, KW., Srivastava, J. Improving bagging performance through multi-algorithm ensembles. Front. Comput. Sci. 6, 498–512 (2012). https://doi.org/10.1007/s11704-012-1163-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11704-012-1163-6

Keywords

Navigation