Abstract
Two algorithms which learn decision trees from examples and their EBL (explanation- based learning) generated rules are presented. The first, IDG-1, learns correct but incomplete trees. It transforms — guided by examples — a rule set into a decision tree which is tailored to efficient execution. Tests done in an example domain show that these trees can be executed much faster than the corresponding EBL generated rule sets even if various methods to optimize rule execution have been applied. Consequently, IDG-1 is one method to ease the utility problem of EBL. The second algorithm, IDG-2, induces complete but no longer entirely correct trees. When compared with trees learned by ID3, the trees induced by IDG-2 showed significantly lower error rates. Since both algorithms construct a tree in a very similar way this demonstrates that the conditions derived from examples and a domain theory via EBL are better suited for tree induction than the simple conditions ID3 constructs from the example descriptions. Both approaches can — under certain conditions — also be used if the rules are generated by other means than EBL.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Davis, R., “Diagnostic reasoning based on structure and behavior“, Artificial Intelligence 24,p. 347–410,1984.
DeJong, G., Mooney, R., “Explanation-based learning: An alternative view“, Machine Learning, Vol. 1, Nr. 2,1986.
Flann, N.S., Dietterich, T.G., “A study of explanation-based methods for inductive learning“, Machine Learning, Vol. 3, Nr. 4,1989.
Forgy, C.L., “Rete: A fast algorithm for the many pattern/many object pattern match problem“, Artificial Intelligence 19, p. 17–37,1982.
Friedrich, G., Nejdl, W., “Increasing the information-theoretic content of diagnostic examples using a domain model“, 9th International workshop expert systems and their applications, Specialized conference Second generation expert systems, Avignon 1989.
Keller, R.M., “Defining operationality for explanation-based learning“, AAAI87.
Matheus, C.J., “Feature construction: An analytic framework and an application to decision trees“, Ph.D. thesis, University of Illinois at Urbana-Champaign, Report No. UIUCDCS-R-89–1559, 1989.
Minton, S., “Quantitative results concerning the utility of explanation-based learning“, AAAI 88.
Mitchell, T.M., Keller, R., Kedar-Cabelli, S.,“Explanation-based generalization: a unifying view“, Machine Learning, Vol. 1, Nr. 1,1986.
Norton, S.W., “Generating better decision trees“, IJCAI 89.
Quinlan, J.R., “Induction of decision trees“, Machine Learning, Vol. 1, Nr. 1,1986.
Resnick, P., “Generalizing on multiple grounds: Performance learning in model-based troubleshooting“, AI-TR 1052, MIT, 1989.
Zercher, K., “Model-based learning of rules for error diagnosis“, in Hoeppner, W. (Ed.), Proceedings of the 12th German workshop on artificial intelligence (GWAI 88), Springer, 1988.
Zercher, K., “Modellbasiertes Lernen von Regeln zur Fehlerdiagnose“, Diplomarbeit, Universität Karlsruhe, 1988.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1990 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zercher, K. (1990). Constructing decision trees from examples and their explanation-based generalizations. In: Marburger, H. (eds) GWAI-90 14th German Workshop on Artificial Intelligence. Informatik-Fachberichte, vol 251. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-76071-6_30
Download citation
DOI: https://doi.org/10.1007/978-3-642-76071-6_30
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-53132-6
Online ISBN: 978-3-642-76071-6
eBook Packages: Springer Book Archive