Skip to main content

Ensemble Classification

  • Chapter
  • First Online:
Principles of Data Mining

Part of the book series: Undergraduate Topics in Computer Science ((UTICS))

  • 3470 Accesses

Abstract

This chapter is concerned with ensemble classification, i.e. using a set of classifiers to classify unseen data rather than just a single one. The classifiers in the ensemble all predict the correct classification of each unseen instance and their predictions are then combined using some form of voting system.

The idea of a random forest of classifiers is introduced and issues relating to the selection of a different training set and/or a different set of attributes from a given dataset when constructing each of the classifiers are discussed.

A number of alternative ways of combining the classifications produced by an ensemble of classifiers are considered. The chapter concludes with a brief discussion of a distributed processing approach to dealing with the large amount of computation often required to generate an ensemble.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 49.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Ho, T. K. (1995). Random decision forests. International Conference on Document Analysis and Recognition, 1, 278.

    Article  Google Scholar 

  2. Breiman, L. (2001). Random forests. Machine Learning, 45(1), 5–32.

    Article  Google Scholar 

  3. Stahl, F., & Bramer, M. (2011). Random prism: an alternative to random forests. In Research and development in intelligent systems XXVIII (pp. 5–18). Springer.

    Chapter  Google Scholar 

  4. Breiman, L. (1996). Bagging predictors. Machine Learning, 24(2), 123–140.

    MATH  Google Scholar 

  5. Stahl, F., May, D., & Bramer, M. (2012). Parallel random prism: a computationally efficient ensemble learner for classification. In Research and development in intelligent systems XXIX. Springer.

    Google Scholar 

  6. Panda, B., Herbach, J. S., Basu, S., & Bayardo, R. J. (2009). Planet: massively parallel learning of tree ensembles with mapreduce. Proceedings of the VLDB Endowment, 2, 1426–1437.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer-Verlag London Ltd., part of Springer Nature

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Bramer, M. (2020). Ensemble Classification. In: Principles of Data Mining. Undergraduate Topics in Computer Science. Springer, London. https://doi.org/10.1007/978-1-4471-7493-6_14

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-7493-6_14

  • Published:

  • Publisher Name: Springer, London

  • Print ISBN: 978-1-4471-7492-9

  • Online ISBN: 978-1-4471-7493-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics