Skip to main content

Combining Models of Approximation with Partial Learning

  • Conference paper
  • First Online:
Algorithmic Learning Theory (ALT 2015)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 9355))

Included in the following conference series:

Abstract

In Gold’s framework of inductive inference, the model of partial learning requires the learner to output exactly one correct index for the target object and only the target object infinitely often. Since infinitely many of the learner’s hypotheses may be incorrect, it is not obvious whether a partial learner can be modified to “approximate” the target object.

Fulk and Jain (Approximate inference and scientific method. Information and Computation 114(2):179–191, 1994) introduced a model of approximate learning of recursive functions. The present work extends their research and solves an open problem of Fulk and Jain by showing that there is a learner which approximates and partially identifies every recursive function by outputting a sequence of hypotheses which, in addition, are also almost all finite variants of the target function.

The subsequent study is dedicated to the question how these findings generalise to the learning of r.e. languages from positive data. Here three variants of approximate learning will be introduced and investigated with respect to the question whether they can be combined with partial learning. Following the line of Fulk and Jain’s research, further investigations provide conditions under which partial language learners can eventually output only finite variants of the target language.

F. Stephan was partially supported by NUS grants R146-000-181-112 and R146-000-184-112; S. Zilles was partially supported by the Natural Sciences and Engineering Research Council of Canada (NSERC).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Angluin, D.: Inductive inference of formal languages from positive data. Information and Control 45(2), 117–135 (1980)

    Article  MathSciNet  MATH  Google Scholar 

  2. Bārzdiņs̆, J.: Two theorems on the limiting synthesis of functions. In: Theory of Algorithms and Programs, vol. 1, pp. 82–88. Latvian State University (1974) (in Russian)

    Google Scholar 

  3. Case, J., Lynes, C.: Machine inductive inference and language identification. In: Nielsen, M., Schmidt, E.M. (eds.) Automata, Languages and Programming. LNCS, vol. 140, pp. 107–115. Springer, Heidelberg (1982)

    Chapter  Google Scholar 

  4. Case, J., Smith, C.: Comparison of identification criteria for machine inductive inference. Theoretical Computer Science 25, 193–220 (1983)

    Article  MathSciNet  MATH  Google Scholar 

  5. Fulk, M., Jain, S.: Approximate inference and scientific method. Information and Computation 114, 179–191 (1994)

    Article  MathSciNet  MATH  Google Scholar 

  6. Gao, Z., Jain, S., Stephan, F.: On conservative learning of recursively enumerable languages. In: Bonizzoni, P., Brattka, V., Löwe, B. (eds.) CiE 2013. LNCS, vol. 7921, pp. 181–190. Springer, Heidelberg (2013)

    Google Scholar 

  7. Gao, Z., Stephan, F., Zilles, S.: Partial learning of recursively enumerable languages. In: Jain, S., Munos, R., Stephan, F., Zeugmann, T. (eds.) ALT 2013. LNCS, vol. 8139, pp. 113–127. Springer, Heidelberg (2013)

    Chapter  Google Scholar 

  8. Gold, E.M.: Language identification in the limit. Information and Control 10, 447–474 (1967)

    Article  MathSciNet  MATH  Google Scholar 

  9. Jain, S., Osherson, D., Royer, J.S., Sharma, A.: Systems that learn: an introduction to learning theory. MIT Press (1999)

    Google Scholar 

  10. Jain, S., Martin, E., Stephan, F.: Learning and classifying. Theoretical Computer Science 482, 73–85 (2013)

    Article  MathSciNet  MATH  Google Scholar 

  11. Martin, E., Osherson, D.N.: Elements of scientific inquiry. MIT Press (1998)

    Google Scholar 

  12. Osherson, D.N., Stob, M., Weinstein, S.: Learning strategies. Information and Control 53, 32–51 (1982)

    Article  MathSciNet  MATH  Google Scholar 

  13. Osherson, D.N., Stob, M., Weinstein, S.: Systems that learn: an introduction to learning theory for cognitive and computer scientists. MIT Press (1986)

    Google Scholar 

  14. Rogers, Jr., H.: Theory of recursive functions and effective computability. MIT Press (1987)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ziyuan Gao .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Gao, Z., Stephan, F., Zilles, S. (2015). Combining Models of Approximation with Partial Learning. In: Chaudhuri, K., GENTILE, C., Zilles, S. (eds) Algorithmic Learning Theory. ALT 2015. Lecture Notes in Computer Science(), vol 9355. Springer, Cham. https://doi.org/10.1007/978-3-319-24486-0_4

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-24486-0_4

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-24485-3

  • Online ISBN: 978-3-319-24486-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics