Abstract
It is well known that any statistical model for supervised or unsupervised classification can be realized as a neural network. This discussion is devoted to supervised classification and therefore the essential framework is the family of feedforward nets.
Ciampi and Lechevallier have studied two- and three-hidden-layer feedforward neural nets that are equivalent to trees, characterized by neurons with “hard” thresholds. Softening the thresholds has led to more general models. Also, neural nets that realize additive models have been studied, as well as networks of networks that represent a “mixed” classifier (predictor) consisting of a tree component and an additive component. Various “dependent” variables have been studied, including the case of censored survival times.
A new development has recently been proposed: the soft tree. A soft tree can be represented as a particular type of hierarchy of experts. This representation can be shown to be equivalent to that of Ciampi and Lechevallier. However, it leads to an appealing interpretation, to other possible generalizations and to a new approach to training. Soft trees for classification and prediction of a continuous variable will be presented. Comparisons between conventional trees (trees with hard thresholds) and soft trees will be discussed and it will be shown that the soft trees achieve better predictions than the hard tree.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Chaudhuri, P., Lo W.-D., Loh, W.-Y., and Yang, C.-C. (1995). Generalized regression trees, Statistica Sinica, 5, 641–666.
Chipman, H., George, E., and McCulloch, R. (2002). Bayesian treed models, Machine Learning, 48, 299–320.
Ciampi, A. (1991), Generalized regression trees, Computational Statistics & Data Analysis, 12, 57–78.
Ciampi, A., Couturier, A., and Li, S. (2002). Prediction trees with soft nodes for binary outcomes, Statistics in Medicine, 21, 1145–1165.
Ciampi, A. and Lechevallier Y. (1995a). Réseaux de neurones et modèles statistiques, La revue de Modulad, 15, 27–46.
Ciampi, A., and Lechevallier, Y. (1995b). Designing neural networks from statistical models: A new approach to data exploration, In Proceedings of the 1st International Conference on Knowledge Discovery and Data Mining, pp. 45–50, AAAI Press, Menlo Park, California.
Ciampi, A., and Lechevallier Y. (1997). Statistical models as building blocks of neural networks, Communication in Statistics, 26, 991–1009.
Ciampi, A., and Lechevallier, Y. (2000). Constructing artificial neural networks for censored survival data from statistical models, In Data Analysis, Classification and Related Methods (Eds., H. Kiers, J.-P. Rasson, P. Groenen, and M. Schader), pp. 223–228, Springer-Verlag, New York.
Ciampi, A., and Lechevallier, Y. (2001). Training an artificial neural network predictor from censored survival data, In 10th International Symposium on Applied Stochastic Models and Data Analysis (Eds., G. Govaerts, J. Janssen, and N. Limnios), Vol. 1, pp. 332–337, Université de Technologie de Compiègne, France.
Ciampi, A., and Zhang F. (2002). A new approach to training back-propagation artificial neural networks: Empirical evaluation on ten data sets from clinical studies, Statistics in Medicine, 21, 1309–1330.
Hastie, T. J., and Tibshirani, R. J. (1990). Generalized Additive Models, Chapman and Hall, New York.
Hastie, T., Tibshirani, R., and Friedman, J. H. (2001). The Elements of Statistical Learning: Data Mining, Inference and Prediction, Springer-Verlag, New York.
Harrell, F. E., Jr. (2002). Regression Modeling Strategies, Springer-Verlag, New York.
Hornik, K., Stinchcombe, M., and White H. (1989). Multilayer feedforward networks are universal approximators, Neural Networks, 2, 359–366.
Jordan, M. I., and Jacobs, R. A. (1994). Hierarchical mixtures of experts and the EM algorithm, Neural Computation, 6, 181–214.
McCullagh, P., and Nelder, J. (1989). Generalized Linear Models, Second edition, Chapman and Hall, London, U.K.
Venable, W. N., and Ripley, B. D. (2002). Modern Applied Statistics with S, Fourth edition, Springer-Verlag, New York.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2007 Birkhäuser Boston
About this chapter
Cite this chapter
Ciampi, A., Lechevallier, Y. (2007). Statistical Models and Artificial Neural Networks: Supervised Classification and Prediction Via Soft Trees. In: Auget, JL., Balakrishnan, N., Mesbah, M., Molenberghs, G. (eds) Advances in Statistical Methods for the Health Sciences. Statistics for Industry and Technology. Birkhäuser Boston. https://doi.org/10.1007/978-0-8176-4542-7_16
Download citation
DOI: https://doi.org/10.1007/978-0-8176-4542-7_16
Publisher Name: Birkhäuser Boston
Print ISBN: 978-0-8176-4368-3
Online ISBN: 978-0-8176-4542-7
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)