Skip to main content

Grammatical inference: An old and new paradigm

  • Conference paper
  • First Online:
Algorithmic Learning Theory (ALT 1995)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 997))

Included in the following conference series:

  • 165 Accesses

Abstract

In this paper, we provide a survey of recent advances in the field “grammatical inference” with a particular emphasis on the results concerning the learnability of target classes represented by deterministic finite automata, context-free grammars, hidden Markov models, stochastic context-free grammars, simple recurrent neural networks, and casebased representations.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. N. Abe and H. Mamitsuka. A new method for predicting protein secondary structures based on stochastic tree grammars. In Proceedings of 11th International Conference on Machine Learning, 1994.

    Google Scholar 

  2. N. Abe and M. K. Warmuth. On the computational complexity of approximating distributions by probabilistic automata. Machine Learning, 9:205–260, 1992.

    Google Scholar 

  3. D. W. Aha, D. Kibler, and M. K. Albert. Instance-based learning algorithms. Machine Learning, 6:37–66, 1991.

    Google Scholar 

  4. A. V. Aho and J. D. Ullman. The Theory of Parsing, Translation and Compiling, Vol. I: Parsing. Prentice Hall, Englewood Cliffs, N.J., 1972.

    Google Scholar 

  5. D. Angluin. Finding patterns common to a set of strings. Journal of Computer and System Sciences, 21:46–62, 1980.

    Article  Google Scholar 

  6. D. Angluin. Inductive inference of formal languages from positive data. Information and Control, 45:117–135, 1980.

    Article  Google Scholar 

  7. D. Angluin. A note on the number of queries needed to identify regular languages. Information and Control, 51:76–87, 1981.

    Article  Google Scholar 

  8. D. Angluin. Inference of reversible languages. Journal of the ACM, 29:741–765, 1982.

    Article  Google Scholar 

  9. D. Angluin. Learning regular sets from queries and counter-examples. Information and Computation, 75:87–106, 1987.

    Article  Google Scholar 

  10. D. Angluin. Queries and concept learning. Machine Learning, 2:319–342, 1988.

    Google Scholar 

  11. D. Angluin. Negative results for equivalence queries. Machine Learning, 5:121–150, 1990.

    Google Scholar 

  12. D. Angluin. Computational learning theory: survey and selected bibliography. In Proceedings of 24th Annual ACM Symposium on Theory of Computing, pages 351–369. ACM Press, 1992.

    Google Scholar 

  13. D. Angluin and M. Kharitonov. When won't membership queries help? In Proceedings of 23rd Annual ACM Symposium on Theory of Computing, pages 444–454. ACM Press, 1991.

    Google Scholar 

  14. D. Angluin and C. H. Smith. Inductive inference: Theory and methods. ACM Computing Surveys, 15(3):237–269, 1983.

    Article  Google Scholar 

  15. S. Arikawa, T. Shinohara, and A. Yamamoto. Elementary formal systems as a unifying framework for language learning. In Proceedings of 2nd Workshop on Computational Learning Theory, pages 312–327. Morgan Kaufmann, 1989.

    Google Scholar 

  16. J. K. Baker. Trainable grammars for speech recognition. Speech Communication Papers for the 97th Meeting of the Acoustical Society of America, pages 547–550, 1979.

    Google Scholar 

  17. A. Brāzma and K. Cerāns. Efficient learning of regular expressions from good examples. In Proceedings of 4th International Workshop on Analogical and Inductive Inference (AII'94), Lecture Notes in Artificial Intelligence 872, pages 76–90. Springer-Verlag, 1994.

    Google Scholar 

  18. A. Burago. Learning structurally reversible context-free grammars from queries and counterexamples in polynomial time. In Proceedings of 7th Workshop on Computational Learning Theory (COLT'93), pages 140–146. ACM Press, 1994.

    Google Scholar 

  19. R. C. Carrasco and J. Oncina, editors. Proceedings of Second International Colloquium on Grammatical Inference (ICGI-94), Lecture Notes in Artificial Intelligence 862. Springer-Verlag, 1994.

    Google Scholar 

  20. S. Crespi-Reghizzi. An effective model for grammar inference. In B. Gilchrist, editor, Information Processing 71, pages 524–529. Elsevier North-Holland, 1972.

    Google Scholar 

  21. P. Dupon. Regular grammatical inference from positive and negative samples by genetic search: the GIG method. In Proceedings of Second International Colloquium on Grammatical Inference (ICGI-94), Lecture Notes in Artificial Intelligence 862, pages 236–245. Springer-Verlag, 1994.

    Google Scholar 

  22. J. L. Elman. Distributed representations, simple recurrent networks, and grammatical structure. Machine Learning, 7:195–225, 1991.

    Google Scholar 

  23. A. F. Fahmy and A. W. Biermann. Synthesis of real time acceptors. Journal of Symbolic Computation, 15:807–842, 1993.

    Google Scholar 

  24. C. L. Giles, G. Z. Sun, H. H. Chen, Y. C. Lee, and D. Chen. Higher order recurrent networks & grammatical inference. In Advances in Neural Information Processing Systems 2, pages 380–387. Morgan Kaufmann, 1990.

    Google Scholar 

  25. E. M. Gold. Language identification in the limit. Information and Control, 10:447–474, 1967.

    Article  Google Scholar 

  26. E. M. Gold. Complexity of automaton identification from given data. Information and Control, 37:302–320, 1978.

    Article  Google Scholar 

  27. M. Golea, M. Matsuoka, and Y. Sakakibara. Unsupervised learning of time-varying probability distributions using two-layer recurrent networks. Unpublished manuscript, 1995.

    Google Scholar 

  28. J. Hertz, A. Krogh, and R. G. Palmer. Introduction to the Theory of Neural Computation. Addison-Wesley, 1991.

    Google Scholar 

  29. L. Hunter. Artificial Intelligence and Molecular Biology. AAAI Press/MIT Press, 1993.

    Google Scholar 

  30. H. Ishizaka. Polynomial time learnability of simple deterministic languages. Machine Learning, 5:151–164, 1990.

    Google Scholar 

  31. K. P. Jantke and S. Lange. Case-based representation and learning of pattern languages. In Proceedings of 4th Workshop on Algorithmic Learning Theory (ALT'93), Lecture Notes in Artificial Intelligence 744, pages 87–100. Springer-Verlag, 1993.

    Google Scholar 

  32. T. Koshiba. Typed pattern languages and their learnability. In Proceedings of 2nd European Conference on Computational Learning Theory (EuroCOLT'95), Lecture Notes in Artificial Intelligence 904, pages 367–379. Springer-Verlag, 1995.

    Google Scholar 

  33. A. Krogh, M. Brown, I. S. Mian, K. Sjölander, and D. Haussler. Hidden Markov models in computational biology. Applications to protein modeling. Journal of Molecular Biology, 235:1501–1531, Feb. 1994.

    PubMed  Google Scholar 

  34. P. D. Laird. A survey of computational learning theory. In R. B. Banerji, editor, Formal Techniques in Artificial Intelligence — A Sourcebook, pages 173–215. Elsevier Science Publishers, 1990.

    Google Scholar 

  35. K. Lari and S. J. Young. The estimation of stochastic context-free grammars using the inside-outside algorithm. Computer Speech and Language, 4:35–56, 1990.

    Article  Google Scholar 

  36. L. S. Levy and A. K. Joshi. Skeletal structural descriptions. Information and Control, 39:192–211, 1978.

    Google Scholar 

  37. E. Mäkinen. On the structural grammatical inference problem for some classes of context-free grammars. Information Processing Letters, 42:193–199, 1992.

    Google Scholar 

  38. L. Miclet. Grammatical inference. In H. Bunke and A. Sanfeliu, editors, Syntactic and Structural Pattern Recognition — Theory and Applications, pages 237–290. World Scientific, 1986.

    Google Scholar 

  39. S. Miyano, A. Shinohara, and T. Shinohara. Which classes of elementary formal systems are polynomial-time learnable? In Proceedings of 2nd Workshop on Algorithmic Learning Theory (ALT'91), pages 139–150. Japanese Society for Artificial Intelligence, Ohmsha, Ltd, 1991.

    Google Scholar 

  40. J. Oncina, P. Garcia, and E. Vidal. Learning subsequential transducers for pattern recognition interpretation tasks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 15:448–458, 1993.

    Article  Google Scholar 

  41. F. Pereira and Y. Schabes. Inside-outside reestimation for partially bracketed corpora. In Proceedings of 30th Annual Meeting of the Association for Computational Linguistics, pages 128–135, 1992.

    Google Scholar 

  42. L. Pitt. Inductive inference, DFAs, and computational complexity. In Proceedings of AII-89 Workshop on Analogical and Inductive Inference (Lecture Notes in Computer Science, 397), pages 18–44. Springer-Verlag, 1989.

    Google Scholar 

  43. L. Pitt and M. K. Warmuth. The minimum consistent DFA problem cannot be approximated within any polynomial. In Proceedings of 21st Annual ACM Symposium on Theory of Computing. ACM Press, 1989.

    Google Scholar 

  44. L. R. Rabiner. A tutorial on hidden Markov models and selected applications in speech recognition. Proc IEEE, 77(2):257–286, 1989.

    Google Scholar 

  45. R. L. Rivest. Learning decision lists. Machine Learning, 2:229–246, 1987.

    Google Scholar 

  46. D. Ron and R. Rubinfeld. Learning fallible deterministic finite automata. Machine Learning, 18:149–185, 1995.

    Google Scholar 

  47. Y. Sakakibara. Learning context-free grammars from structural data in polynomial time. Theoretical Computer Science, 76:223–242, 1990.

    Google Scholar 

  48. Y. Sakakibara. On learning from queries and counterexamples in the presence of noise. Information Processing Letters, 37:279–284, 1991.

    Google Scholar 

  49. Y. Sakakibara. Efficient learning of context-free grammars from positive structural examples. Information and Computation, 97:23–60, 1992.

    Google Scholar 

  50. Y. Sakakibara, M. Brown, R. Hughey, I. S. Mian, K. Sjolander, R. C. Underwood, and D. Haussler. Stochastic context-free grammars for tRNA modeling. Nucleic Acids Research, 22:5112–5120, 1994.

    PubMed  Google Scholar 

  51. Y. Sakakibara and M. Golea. Simple recurrent networks as generalized hidden markov models with distributed representations. Unpublished manuscript, 1995.

    Google Scholar 

  52. Y. Sakakibara, K. P. Jantke, and S. Lange. Learning languages by collecting cases and tuning parameters. In Proceedings of 5th International Workshop on Algorithmic Learning Theory (ALT'94), Lecture Notes in Artificial Intelligence 872, pages 532–546. Springer-Verlag, 1994.

    Google Scholar 

  53. Y. Sakakibara and R. Siromoney. A noise model on learning sets of strings. In Proceedings of 5th Workshop on Computational Learning Theory (COLT'92), pages 295–302. ACM Press, 1992.

    Google Scholar 

  54. D. B. Searls. The linguistics of DNA. American Scientist, 80:579–591, Nov.–Dec. 1992.

    Google Scholar 

  55. T. Shinohara. Inductive inference from positive data is powerful. In Proceedings of 3rd Workshop on Computational Learning Theory, pages 97–110. Morgan Kaufmann, 1990.

    Google Scholar 

  56. A. Stolcke and S. Omohundro. Inducing probabilistic grammars by bayesian model merging. In Proceedings of Second International Colloquium on Grammatical Inference (ICGI-94), Lecture Notes in Artificial Intelligence 862, pages 106–118. Springer-Verlag, 1994.

    Google Scholar 

  57. Y. Takada. Grammatical inference for even linear languages based on control sets. Information Processing Letters, 28:193–199, 1988.

    Article  Google Scholar 

  58. Y. Takada. A hierarchy of language families learnable by regular language learners. In Proceedings of Second International Colloquium on Grammatical Inference (ICGI-94), Lecture Notes in Artificial Intelligence 862, pages 16–24. Springer-Verlag, 1994.

    Google Scholar 

  59. L. G. Valiant. A theory of the learnable. Communications of the ACM, 27:1134–1142, 1984.

    Article  Google Scholar 

  60. T. Yokomori. Polynomial-time learning of very simple grammars from positive data. In Proceedings of 4th Workshop on Computational Learning Theory (COLT'91), pages 213–227. Morgan Kaufmann, 1991.

    Google Scholar 

  61. T. Yokomori. Learning nondeterministic finite automata from queries and counterexamples. In Furukawa, Michie, and Muggleton, editors, Machine Intelligence 13, pages 169–189. Oxford Univ. Press, 1994.

    Google Scholar 

  62. T. Yokomori. On polynomial-time learnability in the limit of strictly deterministic automata. To appear in Machine Learning, 1995.

    Google Scholar 

  63. T. Zeugmann and S. Lange. A guided tour across the boundaries of learning recursive languages. GOSLER-Report 26, TH Leipzig, FB Mathematik und Informatik, 1994.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Klaus P. Jantke Takeshi Shinohara Thomas Zeugmann

Rights and permissions

Reprints and permissions

Copyright information

© 1995 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Sakakibara, Y. (1995). Grammatical inference: An old and new paradigm. In: Jantke, K.P., Shinohara, T., Zeugmann, T. (eds) Algorithmic Learning Theory. ALT 1995. Lecture Notes in Computer Science, vol 997. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-60454-5_25

Download citation

  • DOI: https://doi.org/10.1007/3-540-60454-5_25

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-60454-9

  • Online ISBN: 978-3-540-47470-8

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics