Abstract
Recently, some methods for the induction of Decision Trees have received much theoretical attention. While some of these works focused on efficient top-down induction algorithms, others investigated the pruning of large trees to obtain small and accurate formulae. This paper discusses the practical possibility of combining and generalizing both approaches, to use them on various classes of concept representations, not strictly restricted to decision trees or formulae built from decision trees. The algorithm, WIREI, is able to produce decision trees, decision lists, simple rules, disjunctive normal form formulae, a variant of multilinear polynomials, and more. This shifting ability allows to reduce the risk of deviating from valuable concepts during the induction. As an example, in a previously used simulated noisy dataset, the algorithm managed to find systematically the target concept itself, when using an adequate concept representation. Further experiments on twenty-two readily available datasets show the ability of WIREI to build small and accurate concept representations, which lets the user choose his formalism to best suit his interpretation needs, in particular for mining purposes.
Chapter PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
Blake, C., Keogh, E., Merz, C.J.: UCI repository of machine learning databases (1998), http://www.ics.uci.edu/~mlearn/MLRepository.html
Buntine, W., Niblett, T.: A further comparison of splitting rules for Decision-Tree induction. Machine Learning, 75–85 (1992)
Clark, P., Boswell, R.: Rule induction with CN2: some recent improvements. In: Proc. of the 6 th European Working Session in Learning, pp. 151–161 (1991)
Domingos, P.: A Process-oriented Heuristic for Model selection. In: Proc. of the 15 th International Conference on Machine Learning, pp. 127–135 (1998)
Franck, E., Witten, I.: Using a Permutation Test for Attribute selection in Decision Trees. In: Proc. of the 15th International Conference on Machine Learning, pp. 152–160 (1998)
Holte, R.C.: Very simple classification rules perform well on most commonly used datasets. Machine Learning, 63–91 (1993)
Kearns, M.J., Mansour, Y.: On the boosting ability of top-down decision tree learning algorithms. In: Proceedings of the Twenty-Eighth Annual ACM Symposium on the Theory of Computing, pp. 459–468 (1996)
Kearns, M.J., Mansour, Y.: A Fast, Bottom-up Decision Tree Pruning algorithm with Near-Optimal generalization. In: Proc. of the 15 th International Conference on Machine Learning (1998)
Nock, R., Gascuel, O.: On learning decision committees. In: Proc. of the 12th International Conference on Machine Learning, pp. 413–420 (1995)
Nock, R., Jappy, P.: On the power of decision lists. In: Proc. of the 15th International Conference on Machine Learning, pp. 413–420 (1998)
Nock, R., Jappy, P.: A top-down and prune induction scheme for constrained decision committees. In: Proc. of the 3rd International Symposium on Intelligent Data Analysis (1999) (accepted)
Quinlan, J.R.: Bagging, Boosting and C4.5. In: Proc. of AAAI 1996, pp. 725–730 (1996)
Schapire, R.E., Singer, Y.: Improved boosting algorithms using confidence rated predictions. In: Proceedings of the Eleventh Annual ACM Conference on Computational Learning Theory, pp. 80–91 (1998)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Nock, R., Sebban, M., Jappy, P. (1999). Experiments on a Representation-Independent “Top-Down and Prune” Induction Scheme. In: Żytkow, J.M., Rauch, J. (eds) Principles of Data Mining and Knowledge Discovery. PKDD 1999. Lecture Notes in Computer Science(), vol 1704. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-48247-5_24
Download citation
DOI: https://doi.org/10.1007/978-3-540-48247-5_24
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-66490-1
Online ISBN: 978-3-540-48247-5
eBook Packages: Springer Book Archive