Skip to main content

Max-Cosine Matching Based Neural Models for Recognizing Textual Entailment

  • Conference paper
  • First Online:
Database Systems for Advanced Applications (DASFAA 2017)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 10177))

Included in the following conference series:

Abstract

Recognizing textual entailment is a fundamental task in a variety of text mining or natural language processing applications. This paper proposes a simple neural model for RTE problem. It first matches each word in the hypothesis with its most-similar word in the premise, producing an augmented representation of the hypothesis conditioned on the premise as a sequence of word pairs. The LSTM model is then used to model this augmented sequence, and the final output from the LSTM is fed into a softmax layer to make the prediction. Besides the base model, in order to enhance its performance, we also proposed three techniques: the integration of multiple word-embedding library, bi-way integration, and ensemble based on model averaging. Experimental results on the SNLI dataset have shown that the three techniques are effective in boosting the predicative accuracy and that our method outperforms several state-of-the-state ones.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    http://lasagne.readthedocs.io/en/latest/.

References

  1. Aharon, R.B., Szpektor, I., Dagan, I.: Generating entailment rules from FrameNet. In: Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics (Short Papers), pp. 241–246. Association for Computational Linguistics (2010)

    Google Scholar 

  2. Baker, C.F., Fillmore, C.J., Lowe, J.B.: The Berkeley FrameNet project. In: Proceedings of the 36th Annual Meeting of the Association for Computational Linguistics and 17th International Conference on Computational Linguistics, vol. 1, pp. 86–90. Association for Computational Linguistics (1998)

    Google Scholar 

  3. Bengio, Y., Simard, P., Frasconi, P.: Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 5(2), 157–166 (1994)

    Article  Google Scholar 

  4. Bowman, S.R., Angeli, G., Potts, C., Manning, C.D.: A large annotated corpus for learning natural language inference. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 632–642 (2015)

    Google Scholar 

  5. Bowman, S.R., Gauthier, J., Rastogi, A., Gupta, R., Manning, C.D., Potts, C.: A fast unified model for parsing and sentence understanding. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, pp. 1466–1477 (2016)

    Google Scholar 

  6. Cheng, J., Dong, L., Lapata, M.: Long short-term memory-networks for machine reading. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp. 551–561 (2016)

    Google Scholar 

  7. Dagan, I., Dolan, B., Magnini, B., Roth, D.: Recognizing textual entailment: rational, evaluation and approaches. Nat. Lang. Eng. 15(4), i–xvii (2009)

    Google Scholar 

  8. Dagan, I., Glickman, O.: Probabilistic textual entailment: generic applied modeling of language variability. In: Learning Methods for Text Understanding and Mining, pp. 26–29 (2004)

    Google Scholar 

  9. Fellbaum, C. (ed.): Wordnet: An Electronic Lexical Database. MIT Press, Cambridge (1998)

    Google Scholar 

  10. Giampiccolo, D., Magnini, B., Dagan, I., Dolan, B.: The third pascal recognizing textual entailment challenge. In: Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing, pp. 1–9. Association for Computational Linguistics (2007)

    Google Scholar 

  11. Glickman, O., Dagan, I.: Acquiring lexical paraphrases from a single corpus. In: Recent Advances in Natural Language Processing III, pp. 81–90. John Benjamins Publishing, Amsterdam (2004)

    Google Scholar 

  12. Harris, Z.S.: Distributional structure. Word 10(2–3), 146–162 (1954)

    Article  Google Scholar 

  13. Harabagiu, S., Hickl, A.: Methods for using textual entailment in open-domain question answering. In: Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics, pp. 905–912 (2006)

    Google Scholar 

  14. Hochreiter, S.: Untersuchungen zu dynamischen neuronalen netzen. Technische Universität München, Diploma (1991)

    Google Scholar 

  15. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  16. Kingma, D., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)

  17. Landauer, T.K., Dumais, S.T.: A solution to Platos problem: the latent semantic analysis theory of acquisition, induction, and representation of knowledge. Psychol. Rev. 104(2), 211–240 (1997)

    Article  Google Scholar 

  18. Lin, D., Pantel, P.: DIRT - Discovery of inference rules from text. In: Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 323–328 (2001)

    Google Scholar 

  19. Lloret, E., Ferrández, O., Munoz, R., Palomar, M.: A text summarization approach under the influence of textual entailment. In: Proceedings of the 5th International Workshop on Natural Language Processing and Cognitive Science, pp. 22–31 (2008)

    Google Scholar 

  20. Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems (NIPS), pp. 3111–3119 (2013)

    Google Scholar 

  21. Mou, L., Men, R., Li, G., Xu, Y., Zhang, L., Yan, R., Jin, Z.: Natural language inference by tree-based convolution and heuristic matching. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Short Papers), pp. 130–136 (2016)

    Google Scholar 

  22. Pekar, V.: Acquisition of verb entailment from text. In: Proceedings of the Main Conference on Human Language Technology Conference of the North American Chapter of the Association of Computational Linguistics, pp. 49–56. Association for Computational Linguistics (2006)

    Google Scholar 

  23. Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods on Natural Language Processing, pp. 1532–1543 (2014)

    Google Scholar 

  24. Rocktäschel, T., Grefenstette, E., Hermann, K.M., Kočiský, T., Blunsom, P.: Reasoning about entailment with neural attention. ArXiv Preprint arXiv:1509.06664 (2016)

  25. Sekine, S.: Automatic paraphrase discovery based on context and keywords between NE pairs. In: Proceedings of International Workshop on Paraphrase (IWP 2005), pp. 4–6 (2005)

    Google Scholar 

  26. Srivastava, N.: Improving neural networks with dropout. Ph.D. Thesis, University of Toronto (2013)

    Google Scholar 

  27. Szpektor, I., Dagan, I.: Augmenting wordnet-based inference with argument mapping. In: Proceedings of the 2009 Workshop on Applied Textual Inference, pp. 27–35. Association for Computational Linguistics (2009)

    Google Scholar 

  28. Szpektor, I., Tanev, H., Dagan, I., Coppola, B., Kouylekov, M.: Unsupervised acquisition of entailment relations from the web. Nat. Lang. Eng. 21(01), 3–47 (2015)

    Article  Google Scholar 

  29. Wang, S., Jiang, J.: Learning natural language inference with LSTM. In: Proceedings of the 15th Annual Conference of the North American Chapter of the Association for Computational Linguistics, pp. 700–704 (2016)

    Google Scholar 

  30. Zaremba, W., Sutskever, I., Vinyals, O.: Recurrent neural network regularization. ArXiv Preprint arXiv:1409.2329 (2014)

Download references

Acknowledgments

This work is supported by National High-Tech R&D Program of China (863 Program) (No. 2015AA015404), and Science and Technology Commission of Shanghai Municipality (No. 14511106802). We are grateful to the anonymous reviewers for their valuable comments.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhipeng Xie .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Xie, Z., Hu, J. (2017). Max-Cosine Matching Based Neural Models for Recognizing Textual Entailment. In: Candan, S., Chen, L., Pedersen, T., Chang, L., Hua, W. (eds) Database Systems for Advanced Applications. DASFAA 2017. Lecture Notes in Computer Science(), vol 10177. Springer, Cham. https://doi.org/10.1007/978-3-319-55753-3_19

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-55753-3_19

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-55752-6

  • Online ISBN: 978-3-319-55753-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics