Encyclopedia of Database Systems

2018 Edition
| Editors: Ling Liu, M. Tamer Özsu

Decision Tree Classification

  • Alin DobraEmail author
Reference work entry
DOI: https://doi.org/10.1007/978-1-4614-8265-9_554


Classification tree; Decision tree


Decision tree classifiers are decision trees used for classification. As any other classifier, the decision tree classifiers use values of attributes/features of the data to make a class label (discrete) prediction. Structurally, decision tree classifiers are organized like a decision treein which simple conditions on (usually single) attributes label the edge between an intermediate node and its children. Leaves are labeled by class label predictions. A large number of learning methods have been proposed for decision tree classifiers. Most methods have a tree growing and a pruning phase. The tree growing is recursive and consists in selecting an attribute to split on and actual splitting conditions then recurring on the children until the data corresponding to that path is pure or too small in size. The pruning phase eliminates part of the bottom of the tree that learned noise from the data in order to improve the generalization...

This is a preview of subscription content, log in to check access.

Recommended Reading

  1. 1.
    Agresti A. Categorical data analysis. Wiley; 1990.Google Scholar
  2. 2.
    Breiman L, Friedman JH, Olshen RA, Stone CJ. Classification and regression trees. Belmont: Wadsworth; 1984.zbMATHGoogle Scholar
  3. 3.
    Buntine W. Learning classification trees. Artificial Intelligence frontiers in statistics. London: Chapman & Hall; 1993. p. 182–201.CrossRefGoogle Scholar
  4. 4.
    Cox LA, Qiu Y, Kuehner W. Heuristic least-cost computation of discrete classification functions with uncertain argument values. Ann Oper Res. 1989;21(1):1–30.CrossRefMathSciNetzbMATHGoogle Scholar
  5. 5.
    Frank E. Pruning decision trees and lists. Doctoral dissertation, Department of Computer Science, University of Waikato, Hamilton; 2000.Google Scholar
  6. 6.
    Hyafil L, Rivest RL. Constructing optimal binary decision trees is np-complete. Inf Process Lett. 1976;5(1):15–7.CrossRefMathSciNetzbMATHGoogle Scholar
  7. 7.
    James M. Classification algorithms. New York: Wiley; 1985.zbMATHGoogle Scholar
  8. 8.
    Lim T-S, Loh W-Y, Shih Y-S. An empirical comparison of decision trees and other classification methods Department of Statistics, University of Wisconsin, Madison; 1997. (Technical Report 979).Google Scholar
  9. 9.
    Loh W-Y, Shih Y-S. Split selection methods for classification trees. Stat Sin. 1997;7(4):815.MathSciNetzbMATHGoogle Scholar
  10. 10.
    Murthy SK. Automatic construction of decision trees from data: a multi-disciplinary survey. Data Mining Knowle Discov. 1997;2(4):345–89.Google Scholar
  11. 11.
    Quinlan JR. Induction of decision trees. Mach Learn. 1986;1(1):81–106.Google Scholar
  12. 12.
    Quinlan JR. Learning with continuous classes. In: Proceedings of the 5th Australian Joint Conference on Artificial Intelligence; 1992. p. 343–8.Google Scholar
  13. 13.
    Quinlan JR. C4.5: programs for machine learning. Morgan Kaufman; 1993.Google Scholar
  14. 14.
    Murphy OJ, Mccraw RL. Designing storage efficient decision trees. IEEE Trans Comput. 1991;40(3): 315–9.CrossRefGoogle Scholar

Copyright information

© Springer Science+Business Media, LLC, part of Springer Nature 2018

Authors and Affiliations

  1. 1.University of FloridaGainesvilleUSA

Section editors and affiliations

  • Kyuseok Shim
    • 1
  1. 1.School of Elec. Eng. and Computer ScienceSeoul National Univ.SeoulRepublic of Korea