Abstract
This chapter examines some alternative strategies for selecting attributes at each stage of the TDIDT decision tree generation algorithm and compares the size of the resulting trees for a number of datasets. The risk of obtaining decision trees that are entirely meaningless is highlighted, pointing to the importance of a good choice of attribute selection strategy. One of the most widely used strategies is based on minimising entropy (or equivalently maximising information gain) and this approach is illustrated in detail.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer-Verlag London Ltd., part of Springer Nature
About this chapter
Cite this chapter
Bramer, M. (2020). Decision Tree Induction: Using Entropy for Attribute Selection. In: Principles of Data Mining. Undergraduate Topics in Computer Science. Springer, London. https://doi.org/10.1007/978-1-4471-7493-6_5
Download citation
DOI: https://doi.org/10.1007/978-1-4471-7493-6_5
Published:
Publisher Name: Springer, London
Print ISBN: 978-1-4471-7492-9
Online ISBN: 978-1-4471-7493-6
eBook Packages: Computer ScienceComputer Science (R0)