Skip to main content

AdaBoost

  • Living reference work entry
  • First Online:
Computer Vision
  • 70 Accesses

Synonyms

Adaptive boosting; Discrete AdaBoost

Related Concepts

Definition

The AdaBoost algorithm learns a classifier from data by combining additively a number of weak classifiers. The weak classifiers are incorporated sequentially, one at a time, in order to reduce the empirical exponential classification risk of the combined classifier.

Background

Boosting [1, 2], introduced by Robert Schapire in [3], is a general technique for combining the response of several predictors with limited accuracy into a single, more accurate prediction. AdaBoost is a popular implementation of boosting for binary classification [4]. Soon after its introduction, AdaBoost became one of the most popular learning...

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

References

  1. Friedman JH, Hastie T, Tibshirani R (2000) Additive logistic regression: a statistical view of boosting. Ann Stat 28(2):337–374

    Article  MathSciNet  Google Scholar 

  2. Hastie T, Tibshirani R, Friedman J (2001) The elements of statistical learning. Springer, New York

    Book  Google Scholar 

  3. Schapire RE (1990) The strength of weak learnability. Mach Learn 5(2):197–227

    Google Scholar 

  4. Freund Y, Schapire RE (1996) Experiments with a new boosting algorithm. In: Proceedings of the 13th international conference on machine learning, Bari, pp 148–156

    Google Scholar 

  5. Vapnik V (1995) The nature of statistical learning theory. Springer, New York

    Book  Google Scholar 

  6. Viola P, Jones M (2001) Robust real-time object detection. In: Proceedings of IEEE workshop on statistical and computational theories of vision, Vancouver

    Google Scholar 

  7. Papageorgiou CP, Oren M, Poggio T (1998) A general framework for object detection. In: International conference on computer vision, Bombay, pp 555–562

    Google Scholar 

  8. Sun Y, Li J, Hager W (2004) Two new regularized adaboost algorithms. In: Machine learning and applications, Louisville, pp 41–48

    Google Scholar 

  9. Schapire RE, Singer Y (1998) Improved boosting algorithms using confidence-rated predictions. In: Computational learning theory. Springer, New York, pp 80–91

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Paolo Favaro .

Section Editor information

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this entry

Check for updates. Verify currency and authenticity via CrossMark

Cite this entry

Favaro, P., Vedaldi, A. (2021). AdaBoost. In: Computer Vision. Springer, Cham. https://doi.org/10.1007/978-3-030-03243-2_663-1

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-03243-2_663-1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-03243-2

  • Online ISBN: 978-3-030-03243-2

  • eBook Packages: Springer Reference Computer SciencesReference Module Computer Science and Engineering

Publish with us

Policies and ethics