Skip to main content

Statistical Machine Translation Context Modelling with Recurrent Neural Network and LDA

  • Conference paper
  • First Online:
Proceedings of the International Conference on Advanced Intelligent Systems and Informatics 2016 (AISI 2016)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 533))

  • 2559 Accesses

Abstract

Machine Translation of text is a fundamental problem in machine learning that resists solutions that do not take into account the dependencies between words and sentences. Recurrent Neural Networks have recently delivered outstanding results in learning about sequential dependencies in many languages. Arabic language as a target language has not received enough attention in the recent language model experiments due to its, structural and semantic difficulties. In this paper, we present a Statistical Machine Translation (SMT) Context Modelling using Recurrent Neural Networks (RNNs) and Latent Dirichlet Allocation (LDA). This research is based on the state-of-the-art RNN language model by Mikolov. Our preliminary contribution is in integrating and presenting a new hybridization to utilize Recurrent Neural Network sequential word learning dependencies as well as Latent Dirichlet Allocation context and topic classification ability to produce the most accurate language scoring.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Callison-Burch, C., Talbot, D., Osborne, M.: Statistical machine translation with word-and sentence-aligned parallel corpora. In: Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics, p. 175 (2004)

    Google Scholar 

  2. Durrani, N., Fraser, A., Schmid, H.: Model with minimal translation units, but decode with phrases. In: Proceedings of NAACL-HLT, 9–14 June 2013, Atlanta, Georgia (2013)

    Google Scholar 

  3. Brown, P., de Souza, P., Mercer, R., Pietra, V., Lai, J.: Class-based n-gram models of natural language. Computational Linguistics. Comput. Linguist. 18(4), 467–479 (1992)

    Google Scholar 

  4. Lipton, Z., Berkowitz, J., Elkan, C.: A Critical review of recurrent neural networks for sequence learning’, arXiv preprint arXiv:1506.00019 (2015)

  5. Farabet, C., Couprie, C., Najman, L., LeCun, Y.: Learning hierarchical features for scene labeling. IEEE Trans. Pattern Anal. Mach. Intell. 35(8), 1915–1929 (2013)

    Article  Google Scholar 

  6. Zhao, B., Tam, Y.: Bilingual recurrent neural networks for improved statistical machine translation. In: Spoken Language Technology Workshop (SLT), 7–10 December 2014, South Lake Tahoe, NV. IEEE (2014)

    Google Scholar 

  7. Sundermeyer, M., Alkhouli, T., Wuebker, J., Ney, H.: Translation modeling with bidirectional recurrent neural networks. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), 14–25 October 2014, Doha, Qatar (2014)

    Google Scholar 

  8. Hopfield, J.J.: Neural networks and physical systems with emergent collective computational abilities. Proc. Nat. Acad. Sci. 79(8), 2554–2558 (1982)

    Article  MathSciNet  Google Scholar 

  9. Jordan, M.: Serial order: a parallel distributed processing approach. Technical report 8604, Institute for Cognitive Science, University of California, San Diego (1986)

    Google Scholar 

  10. Elman, J.: Finding structure in time. Cogn. Sci. 14, 179–211 (1990)

    Article  Google Scholar 

  11. Mikolov, T., Zweig, G.: Context dependent recurrent neural network language model. In: 2012 workshop on Spoken Language Technology, pp. 234–239 (2012)

    Google Scholar 

  12. Kombrink, S., Mikolov, T., Karafiat, M., Burget, L.: Recurrent neural network based language model. In: INTERSPEECH 2010, 11th Annual Conference of the International Speech Communication Association, Makuhari, Chiba, Japan, 26–30 September 2010, pp. 1045–1048 (2010)

    Google Scholar 

  13. Kalchbrenner, N., Blunsom, P.: Recurrent continuous translation models. In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, October 2013, Seattle, Washington, USA, pp. 1700–1709 (2013)

    Google Scholar 

  14. Liu, S., Yang, N., Li, Zhou, M.: A recursive recurrent neural network for statistical machine translation. In: Proceedings of ACL, pp. 1491–1550 (2014)

    Google Scholar 

  15. Hu, Y., Auli, M., Gao, Q., Gao, J.: Minimum translation modeling with recurrent neural networks. In: Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics, April 2014

    Google Scholar 

  16. Schuster, M., Paliwal, K.: Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 45(11), 2673–2681 (1997)

    Article  Google Scholar 

  17. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  18. Schwenk, H.: Continuous space translation models for phrase-based statistical machine translation. In: 25th International Conference on Computational Linguistics (COLING), December, Mumbai, India, pp. 1071–1080 (2012)

    Google Scholar 

  19. Mikolov, T.: Statistical Language Models Based on Neural Networks. Ph.D., Brno University of Technology (2012)

    Google Scholar 

  20. Arabic Gigaword Corpus. https://catalog.ldc.upenn.edu/LDC2011T11

  21. Mikolov, T.: RNNLM Toolkit (2012). http://www.rnnlm.org/. Accessed: 28 Nov 2015

  22. Guessabi, F.: The cultural problems in translating a novel from arabic to english language. AWEJ Special Issue on Translation (2), 224–232 (2013)

    Google Scholar 

  23. https://ar.wikipedia.org/wiki/. Accessed: 20 Dec 2015

  24. Ponweiser, M.: Latent dirichlet allocation in R. Diploma Thesis, Institute for Statistics and Mathematics 2 May 2012

    Google Scholar 

  25. Zhengxian, G., Guodong, Z.: Employing topic modeling for statistical machine translation. In: 2011 IEEE International Conference on Computer Science and Automation Engineering, CSAE (2011)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shrooq Alsenan .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Alsenan, S., Ykhlef, M. (2017). Statistical Machine Translation Context Modelling with Recurrent Neural Network and LDA. In: Hassanien, A., Shaalan, K., Gaber, T., Azar, A., Tolba, M. (eds) Proceedings of the International Conference on Advanced Intelligent Systems and Informatics 2016. AISI 2016. Advances in Intelligent Systems and Computing, vol 533. Springer, Cham. https://doi.org/10.1007/978-3-319-48308-5_8

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-48308-5_8

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-48307-8

  • Online ISBN: 978-3-319-48308-5

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics