Skip to main content

Statistical Models and Artificial Neural Networks: Supervised Classification and Prediction Via Soft Trees

  • Chapter

Part of the book series: Statistics for Industry and Technology ((SIT))

Abstract

It is well known that any statistical model for supervised or unsupervised classification can be realized as a neural network. This discussion is devoted to supervised classification and therefore the essential framework is the family of feedforward nets.

Ciampi and Lechevallier have studied two- and three-hidden-layer feedforward neural nets that are equivalent to trees, characterized by neurons with “hard” thresholds. Softening the thresholds has led to more general models. Also, neural nets that realize additive models have been studied, as well as networks of networks that represent a “mixed” classifier (predictor) consisting of a tree component and an additive component. Various “dependent” variables have been studied, including the case of censored survival times.

A new development has recently been proposed: the soft tree. A soft tree can be represented as a particular type of hierarchy of experts. This representation can be shown to be equivalent to that of Ciampi and Lechevallier. However, it leads to an appealing interpretation, to other possible generalizations and to a new approach to training. Soft trees for classification and prediction of a continuous variable will be presented. Comparisons between conventional trees (trees with hard thresholds) and soft trees will be discussed and it will be shown that the soft trees achieve better predictions than the hard tree.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   149.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD   199.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Chaudhuri, P., Lo W.-D., Loh, W.-Y., and Yang, C.-C. (1995). Generalized regression trees, Statistica Sinica, 5, 641–666.

    MATH  MathSciNet  Google Scholar 

  2. Chipman, H., George, E., and McCulloch, R. (2002). Bayesian treed models, Machine Learning, 48, 299–320.

    Article  MATH  Google Scholar 

  3. Ciampi, A. (1991), Generalized regression trees, Computational Statistics & Data Analysis, 12, 57–78.

    Article  MATH  MathSciNet  Google Scholar 

  4. Ciampi, A., Couturier, A., and Li, S. (2002). Prediction trees with soft nodes for binary outcomes, Statistics in Medicine, 21, 1145–1165.

    Article  Google Scholar 

  5. Ciampi, A. and Lechevallier Y. (1995a). Réseaux de neurones et modèles statistiques, La revue de Modulad, 15, 27–46.

    Google Scholar 

  6. Ciampi, A., and Lechevallier, Y. (1995b). Designing neural networks from statistical models: A new approach to data exploration, In Proceedings of the 1st International Conference on Knowledge Discovery and Data Mining, pp. 45–50, AAAI Press, Menlo Park, California.

    Google Scholar 

  7. Ciampi, A., and Lechevallier Y. (1997). Statistical models as building blocks of neural networks, Communication in Statistics, 26, 991–1009.

    Article  MATH  MathSciNet  Google Scholar 

  8. Ciampi, A., and Lechevallier, Y. (2000). Constructing artificial neural networks for censored survival data from statistical models, In Data Analysis, Classification and Related Methods (Eds., H. Kiers, J.-P. Rasson, P. Groenen, and M. Schader), pp. 223–228, Springer-Verlag, New York.

    Google Scholar 

  9. Ciampi, A., and Lechevallier, Y. (2001). Training an artificial neural network predictor from censored survival data, In 10th International Symposium on Applied Stochastic Models and Data Analysis (Eds., G. Govaerts, J. Janssen, and N. Limnios), Vol. 1, pp. 332–337, Université de Technologie de Compiègne, France.

    Google Scholar 

  10. Ciampi, A., and Zhang F. (2002). A new approach to training back-propagation artificial neural networks: Empirical evaluation on ten data sets from clinical studies, Statistics in Medicine, 21, 1309–1330.

    Article  Google Scholar 

  11. Hastie, T. J., and Tibshirani, R. J. (1990). Generalized Additive Models, Chapman and Hall, New York.

    MATH  Google Scholar 

  12. Hastie, T., Tibshirani, R., and Friedman, J. H. (2001). The Elements of Statistical Learning: Data Mining, Inference and Prediction, Springer-Verlag, New York.

    MATH  Google Scholar 

  13. Harrell, F. E., Jr. (2002). Regression Modeling Strategies, Springer-Verlag, New York.

    Google Scholar 

  14. Hornik, K., Stinchcombe, M., and White H. (1989). Multilayer feedforward networks are universal approximators, Neural Networks, 2, 359–366.

    Article  Google Scholar 

  15. Jordan, M. I., and Jacobs, R. A. (1994). Hierarchical mixtures of experts and the EM algorithm, Neural Computation, 6, 181–214.

    Article  Google Scholar 

  16. McCullagh, P., and Nelder, J. (1989). Generalized Linear Models, Second edition, Chapman and Hall, London, U.K.

    MATH  Google Scholar 

  17. Venable, W. N., and Ripley, B. D. (2002). Modern Applied Statistics with S, Fourth edition, Springer-Verlag, New York.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2007 Birkhäuser Boston

About this chapter

Cite this chapter

Ciampi, A., Lechevallier, Y. (2007). Statistical Models and Artificial Neural Networks: Supervised Classification and Prediction Via Soft Trees. In: Auget, JL., Balakrishnan, N., Mesbah, M., Molenberghs, G. (eds) Advances in Statistical Methods for the Health Sciences. Statistics for Industry and Technology. Birkhäuser Boston. https://doi.org/10.1007/978-0-8176-4542-7_16

Download citation

Publish with us

Policies and ethics