Skip to main content

Active Learning Theory

  • Reference work entry
Encyclopedia of Machine Learning
  • 895 Accesses

Definition

The term active learning applies to a wide range of situations in which a learner is able to exert some control over its source of data. For instance, when fitting a regression function, the learner may itself supply a set of data points at which to measure response values, in the hope of reducing the variance of its estimate. Such problems have been studied for many decades under the rubric of experimental design (Chernoff, 1972; Fedorov, 1972). More recently, there has been substantial interest within the machine learning community in the specific task of actively learning binary classifiers. This task presents several fundamental statistical and algorithmic challenges, and an understanding of its mathematical underpinnings is only gradually emerging. This brief survey will describe some of the progress that has been made so far.

Learning from Labeled and Unlabeled Data

In the machine learning literature, the task of learning a classifier has traditionally been studied in...

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Recommended Reading

  • Angluin, D. (2001). Queries revisited. In Proceedings of the 12th international conference on algorithmic learning theory (pp. 12–31).

    Google Scholar 

  • Balcan, M.-F., Beygelzimer, A., & Langford, J. (2006). Agnostic active learning. In International Conference on Machine Learning (pp. 65–72). New York: ACM Press.

    Google Scholar 

  • Balcan, M.-F., Broder, A., & Zhang, T. (2007). Margin based active learning. In Conference on Learning Theory. pp. 35–50.

    Google Scholar 

  • Baum, E. B., & Lang, K. (1992). Query learning can work poorly when a human oracle is used. In International Joint Conference on Neural Networks.

    Google Scholar 

  • Beygelzimer, A., Dasgupta, S., & Langford, J. (2009). Importance weighted active learning. In International Conference on Machine Learning (pp. 49–56). New York: ACM Press.

    Google Scholar 

  • Cesa-Bianchi, N., Gentile, C., & Zaniboni, L. (2004). Worst-case analysis of selective sampling for linear-threshold algorithms. Advances in Neural Information Processing Systems.

    Google Scholar 

  • Chernoff, H. (1972). Sequential analysis and optimal design. In CBMS-NSF Regional Conference Series in Applied Mathema- tics 8. SIAM.

    Google Scholar 

  • Cohn, D., Atlas, L., & Ladner, R. (1994). Improving generalization with active learning. Machine Learning, 15(2),201–221.

    Google Scholar 

  • Dasgupta, S. (2005). Coarse sample complexity bounds for active learning. Advances in Neural Information Processing Systems.

    Google Scholar 

  • Dasgupta, S., Kalai, A., & Monteleoni, C. (2005). Analysis of perceptron-based active learning. In 18th Annual Conference on Learning Theory. pp. 249–263.

    Google Scholar 

  • Dasgupta, S., Hsu, D. J., & Monteleoni, C. (2007). A general agnostic active learning algorithm. Advances in Neural Information Processing Systems.

    Google Scholar 

  • Fedorov, V. V. (1972). Theory of optimal experiments. (W. J. Studden & E. M. Klimko, Trans.). New York: Academic Press.

    Google Scholar 

  • Freund, Y., Seung, S., Shamir, E., & Tishby, N. (1997). Selective sampling using the query by committee algorithm. Machine Learning Journal, 28,133–168.

    MATH  Google Scholar 

  • Friedman, E. (2009). Active learning for smooth problems. In Conference on Learning Theory. pp. 343–352.

    Google Scholar 

  • Gilad-Bachrach, R., Navot, A., & Tishby, N. (2005). Query by committeee made real. Advances in Neural Information Processing Systems.

    Google Scholar 

  • Hanneke, S. (2007a). Teaching dimension and the complexity of active learning. In Conference on Learning Theory. pp. 66–81.

    Google Scholar 

  • Hanneke, S. (2007b). A bound on the label complexity of agnostic active learning. In International Conference on Machine Learning. pp. 353–360.

    Google Scholar 

  • Haussler, D. (1992). Decision-theoretic generalizations of the PAC model for neural net and other learning applications. Information and Computation, 100(1),78–150.

    MATH  MathSciNet  Google Scholar 

  • Seung, H. S., Opper, M., & Sompolinsky, H. (1992). Query by committee. In Conference on Computational Learning Theory, pp. 287–294.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Springer Science+Business Media, LLC

About this entry

Cite this entry

Dasgupta, S. (2011). Active Learning Theory. In: Sammut, C., Webb, G.I. (eds) Encyclopedia of Machine Learning. Springer, Boston, MA. https://doi.org/10.1007/978-0-387-30164-8_7

Download citation

Publish with us

Policies and ethics