Abstract
Two algorithms for supervised learning of classifications are discussed from the point of view of the usefulness of including statistical methods. It will be demonstrated that statistical considerations of very general nature (i.e. without assumptions on class distributions) can lead to substantial improvements of the learning procedure and the constructed classifiers. The decision tree learner CAL5 converts real-valued attributes into discrete-valued ones the number of which is not restricted to two. Pruning occurs during tree construction. The hybrid (statistical/neural) algorithm DIPOL solves the problem of choosing the initial architecture and initial weights by statistical methods and replaces additional hidden layers by a Boolean decision function. Both algorithms are also discussed within the framework of the ESPRIT-Project StatLog where about 20 of the most important procedures for classification learning are compared using statistical criteria.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Michie, D., Spiegelhalter, D.J. and Taylor, C.C. (Eds.): Machine Learning, Neural and Statistical Classification, Ellis Norwood, New York 1994.
Minsky, M. and Papert, S.: Perceptrons, MIT Press, Cambridge, MA, 1969.
Müller, W., and Wysotzki, F.: Automatic Construction of Decision Trees for Classification, Annals of Operations Research, 52 (1994), pp. 231–247.
Müller, W., Wysotzki, F.: The Decision Tree Algorithm CAL5 Based on a Statistical Approach to its Splitting Algorithm, in: Nakhaeizadeh, G. and Taylor, C.C. (Eds.), Machine Learning and Statistics, The Interface. John Wiley Sons, New York, 1996.
Nilsson, N.: Learning Machines. McGraw-Hill, New York 1965.
Schulmeister, B. and Wysotzki, F.: DIPOL - A Hybrid Piecewise Linear Classifier, in: Nakhaeizadeh, G. and Taylor, C.C. (Eds.), Machine Learning and Statistics, The Interface. John Wiley Sons, New York 1996.
Unger, S. and Wysotzki, F.: Lernfähige Klassifizierungs- systeme, Akademieverlag, Berlin 1981.
Widrow, B.: Generalization and Information Storage in Networks of ADALINE “Neurons”, in: Self Organizing Systems ( M. C. Yovits, G. T. Jacoby, and G. D. Goldstein, eds), Spartan Books, Washington, D.C. 1962, pp. 435–461.
Werk, R.: Untersuchungen zur Korrektur und zum inkrementellen Lernen von Klassifikationen durch den hybriden Algorithmus DIPOL, Diplomarbeit, Dept. of Computer Science, Technical University of Berlin 1996.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1997 Springer-Verlag Wien
About this paper
Cite this paper
Wysotzki, F., Müller, W., Schulmeister, B. (1997). Automatic Construction of Decision Trees and Neural Nets for Classification Using Statistical Considerations. In: Della Riccia, G., Lenz, HJ., Kruse, R. (eds) Learning, Networks and Statistics. International Centre for Mechanical Sciences, vol 382. Springer, Vienna. https://doi.org/10.1007/978-3-7091-2668-4_8
Download citation
DOI: https://doi.org/10.1007/978-3-7091-2668-4_8
Publisher Name: Springer, Vienna
Print ISBN: 978-3-211-82910-3
Online ISBN: 978-3-7091-2668-4
eBook Packages: Springer Book Archive