Skip to main content

Active Learning in the Non-realizable Case

  • Conference paper
Algorithmic Learning Theory (ALT 2006)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 4264))

Included in the following conference series:

Abstract

Most of the existing active learning algorithms are based on the realizability assumption: The learner’s hypothesis class is assumed to contain a target function that perfectly classifies all training and test examples. This assumption can hardly ever be justified in practice. In this paper, we study how relaxing the realizability assumption affects the sample complexity of active learning. First, we extend existing results on query learning to show that any active learning algorithm for the realizable case can be transformed to tolerate random bounded rate class noise. Thus, bounded rate class noise adds little extra complications to active learning, and in particular exponential label complexity savings over passive learning are still possible. However, it is questionable whether this noise model is any more realistic in practice than assuming no noise at all.

Our second result shows that if we move to the truly non-realizable model of statistical learning theory, then the label complexity of active learning has the same dependence Ω(1/ε 2) on the accuracy parameter ε as the passive learning label complexity. More specifically, we show that under the assumption that the best classifier in the learner’s hypothesis class has generalization error at most β>0, the label complexity of active learning is Ω(β 2/ε 2log(1/δ)), where the accuracy parameter ε measures how close to optimal within the hypothesis class the active learner has to get and δ is the confidence parameter. The implication of this lower bound is that exponential savings should not be expected in realistic models of active learning, and thus the label complexity goals in active learning should be refined.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Freund, Y., Seung, H.S., Shamir, E., Tishby, N.: Selective sampling using the query by committee algorithm. Machine Learning 28(2-3), 133–168 (1997)

    Article  MATH  Google Scholar 

  2. Dasgupta, S., Kalai, A.T., Monteleoni, C.: Analysis of perceptron-based active learning. In: Auer, P., Meir, R. (eds.) COLT 2005. LNCS, vol. 3559, pp. 249–263. Springer, Heidelberg (2005)

    Chapter  Google Scholar 

  3. Dasgupta, S.: Coarse sample complexity bounds for active learning. In: NIPS 2005 (2005)

    Google Scholar 

  4. Balcan, N., Beygelzimer, A., Langford, J.: Agnostic active learning. In: ICML (accepted, 2006)

    Google Scholar 

  5. Tong, S., Koller, D.: Support vector machine active learning with applications to text classification. Journal of Machine Learning Research 2, 45–66 (2002)

    Article  MATH  Google Scholar 

  6. Angluin, D., Laird, P.: Learning from noisy examples. Machine Learning 2(4), 343–370 (1987)

    Google Scholar 

  7. Sakakibara, Y.: On learning from queries and counterexamples in the presence of noise. Information Processing Letters 37(5), 279–284 (1991)

    Article  MATH  MathSciNet  Google Scholar 

  8. Vapnik, V.N.: Estimation of Dependencies Based on Empirical Data. Springer, Heidelberg (1982)

    Google Scholar 

  9. Gentile, C., Helmbold, D.P.: Improved lower bounds for learning from noisy examples: an information-theoretic approach. In: COLT 1998, pp. 104–115. ACM Press, New York (1998)

    Chapter  Google Scholar 

  10. Domingo, C., Gavaldá, R., Watanabe, O.: Adaptive sampling methods for scaling up knowledge discovery algorithms. In: Arikawa, S., Furukawa, K. (eds.) DS 1999. LNCS, vol. 1721, pp. 172–183. Springer, Heidelberg (1999)

    Chapter  Google Scholar 

  11. Castro, R.: Personal communication (March 2006)

    Google Scholar 

  12. Silvey, S.D.: Optimal Design. Chapman and Hall, London (1980)

    MATH  Google Scholar 

  13. Elfving, G.: Selection of nonrepeatable observations for estimation. In: Proceedings of the 3rd Berkeley Symposium on Mathematical Statistics and Probability, vol. 1, pp. 69–75 (1956)

    Google Scholar 

  14. Canetti, R., Even, G., Goldreich, O.: Lower bounds for sampling algorithms for estimating the average. Information Processing Letters 53(1), 17–25 (1995)

    Article  MATH  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2006 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Kääriäinen, M. (2006). Active Learning in the Non-realizable Case. In: Balcázar, J.L., Long, P.M., Stephan, F. (eds) Algorithmic Learning Theory. ALT 2006. Lecture Notes in Computer Science(), vol 4264. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11894841_9

Download citation

  • DOI: https://doi.org/10.1007/11894841_9

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-46649-9

  • Online ISBN: 978-3-540-46650-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics