Skip to main content

Attention-Based Bidirectional Long Short-Term Memory Neural Network for Short Answer Scoring

  • Conference paper
  • First Online:
Machine Learning and Intelligent Communications (MLICOM 2020)

Abstract

The automatic short answer scoring by using computational approaches has been considered the best way to release the workload of human answer raters. In this paper, we designed a novel neural network architecture which is attention-based bidirectional long short-term memory to implement the task of automatic short answer scoring. We evaluate our approach on the Kaggle Short Answer dataset (ASAP-SAS). Our experiment results indicate that our model can scoring short answers more accurately in terms of the quality of the results. Meanwhile, our experiment results demonstrate that our model is more effective and efficient than other baseline methods in most cases.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Dikli, S.: An overview of automated scoring of essays. J. Technol. Learn. Assess. 5(1), 1–35 (2006)

    Google Scholar 

  2. Page, E.B.: The imminence of grading essays by computer. Phi Delta Kappan 48, 238–243 (1966)

    Google Scholar 

  3. Claudia, L., Martin, C.: C-rater: Automated scoring of short-answer questions. Comput. Humanit. 37(4), 389–405 (2003)

    Article  Google Scholar 

  4. Deerwester, S., Dumais, S.T., Furnas, G.W., Landauer, T.K., Harshman, R.: Indexing by latent semantic analysis. J. Am. Soc. Inf. Sci. 41(6), 391–407 (1990)

    Google Scholar 

  5. Landauer, T., Laham, D., Foltz, P.: Automated scoring and annotation of essays with the intelligent essay assessor. In: Automated Essay Scoring: A Cross-Disciplinary Perspective, pp. 87–112 (2003)

    Google Scholar 

  6. Hofmann T.: Probabilistic latent semantic indexing. In: Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 50–57. Association for Computing Machinery ACM, Berkeley (1999)

    Google Scholar 

  7. Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent Dirichlet allocation. J. Mach. Learn. Res. 3, 993–1022 (2003)

    MATH  Google Scholar 

  8. McNamara, D., Crossley, S.A., Mccarthy, P.M.: Linguistic features of writing quality. Written Commun. 27(1), 57–86 (2010)

    Article  Google Scholar 

  9. Gomaa, W.H., Fahmy, A.A., Ans2vec: a scoring system for short answers. In: Hassanien, A., Azar, A., Gaber, T., Bhatnagar, R., F. Tolba, M. (eds) AMLTA 2019, vol. 821, pp. 586–595. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-14118-9_59

  10. Tang, D.: Sentiment-specific representation learning for document-level sentiment analysis. In: Proceedings of the Eighth ACM International Conference on Web Search and Data Mining, pp. 447–452. Association for Computing Machinery (ACM), Shanghai (2015)

    Google Scholar 

  11. Pang, B., Lee, L.: Seeing stars: exploiting class relationships for sentiment categorization with respect to rating scales. In: Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics, pp. 115–124. Association for Computational Linguistics (ACL), Ann Arbor (2005)

    Google Scholar 

  12. Lee, K., Han, S., Myaeng, S.-H.: A discourse-aware neural network-based text model for document-level text classification. J. Inf. Sci. 44(6), 715–735 (2018)

    Article  Google Scholar 

  13. Mikolov T., Chen K., Corrado G., Dean J.: Efficient estimation of word representations in vector space. arXiv:1301.3781[cs.CL], 1–12 (2013)

  14. Collobert, R., Weston, J., Bottou, L., Karlen, M., Kavukcuoglu, K., Kuksa, P.: Natural language processing (almost) from scratch. J. Mach. Learn. Res. 12, 2493–2537 (2011)

    MATH  Google Scholar 

  15. Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pp. 1532–1543. Association for Computational Linguistics (ACL), Doha (2014)

    Google Scholar 

  16. Zhang, H., Litman, D.: Co-attention based neural network for source-dependent essay scoring. In: Proceedings of the Thirteenth Workshop on Innovative Use of NLP for Building Educational Applications, pp. 399–409. Association for Computational Linguistics (ACL), New Orleans (2018)

    Google Scholar 

  17. Ali, M.N.A., Tan, G.Z., Hussain, A.: Bidirectional recurrent neural network approach for Arabic named entity recognition. Future Internet 10(12), 123 (2018)

    Article  Google Scholar 

  18. Alikaniotis, D., Yannakoudakis, H., Rei, M.: Automatic text scoring using neural networks. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, pp. 7–12. Association for Computational Linguistics (ACL), Berlin (2016)

    Google Scholar 

  19. Kim, Y.: Convolutional neural networks for sentence classification. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 1746–1751. Association for Computational Linguistics (ACL), Doha (2014)

    Google Scholar 

  20. Liao, S., Wang, J., Yu, R., Sato, K., Cheng, Z.: CNN for situations understanding based on sentiment analysis of twitter data. In: Proceedings of the 8th International Conference on Advances in Information Technology, Elsevier B.V., pp. 376–381. Macau (2016)

    Google Scholar 

  21. Zhang, Y., Wallace, B.C.: A sensitivity analysis of (and practitioners’ guide to) convolutional neural networks for sentence classification. arXiv:1510.03820[cs.CL], pp. 1–18 (2016)

  22. Zhang, Y., Er, M.J., Venkatesan, R., Wang, N., Pratama, M.: Sentiment classification using comprehensive attention recurrent models. In: Proceedings of the International Joint Conference on Neural Networks, pp. 1562–1569. IEEE, Vancouver (2016)

    Google Scholar 

  23. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  24. Ran, X., Shan, Z., Fang, Y., Lin, C.: An LSTM-based method with attention mechanism for travel time prediction. Sensors 19(4), 861 (2019)

    Article  Google Scholar 

  25. Nowak, J., Taspinar, A., Scherer, R.: LSTM recurrent neural networks for short text and sentiment classification. In: Proceedings of the 16th International Conference on Artificial Intelligence and Soft Computing, pp. 553–562. Springer Verlag, Zakopane (2017)

    Google Scholar 

  26. Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 18(5–6), 602–610 (2005)

    Article  Google Scholar 

  27. Bin, Y., Yang, Y., Shen, F., Xie, N., Shen, T., Li, X.: Describing video with attention-based bidirectional LSTM. IEEE Trans. Cybern. 49(7), 2631–2641 (2019)

    Article  Google Scholar 

  28. Luong, M.-T., Pham, H., Manning, C.D.: Effective approaches to attention-based neural machine translation. In: Proceedings of Conference on Empirical Methods in Natural Language Processing, pp. 1412–1421. Association for Computational Linguistics (ACL), Lisbon (2015)

    Google Scholar 

  29. Yin, W., Ebert, S., Schütze, H.: Attention-based convolutional neural network for machine comprehension. In: Proceedings of the Workshop on Human-Computer Question Answering, pp. 15–21. Association for Computational Linguistics (ACL), San Diego (2016)

    Google Scholar 

  30. Zhang, Y., Shah, R., Chi, M.: Deep learning + student modeling + clustering: a recipe for effective automatic short answer grading. In: Proceedings of the 9th International Conference on Educational Data Mining, pp. 562–567. International Educational Data Mining Society (IEDMS), Raleigh (2016)

    Google Scholar 

  31. Zhang, X., LeCun, Y.: Text Understanding from Scratch. arXiv:1502.01710 [cs.LG] (2016)

  32. Walia, T.S., Josan, G.S., Singh, A.: An efficient automated answer scoring system for Punjabi language. Egyptian Inf. J. 20, 89–96 (2019)

    Article  Google Scholar 

  33. Surya, K., Ekansh, G., Nallakaruppan, K.: Deep learning for short answer scoring. Int. J. Recent Technol. Eng. 7(6), 1712–1715 (2019)

    Google Scholar 

Download references

Acknowledgement

This work is supported by Engineering Applications of Artificial Intelligence Technology Laboratory of Shenzhen Institute of Information Technology (Number: PT201701), the Guangdong Province higher vocational colleges & schools Pearl River scholar funded scheme (2016), and The Scientific and Technological Projects of Shenzhen (No. JCYJ20190808093001772).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Linzhong Xia .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Xia, L., Guan, M., Liu, J., Cao, X., Luo, D. (2021). Attention-Based Bidirectional Long Short-Term Memory Neural Network for Short Answer Scoring. In: Guan, M., Na, Z. (eds) Machine Learning and Intelligent Communications. MLICOM 2020. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 342. Springer, Cham. https://doi.org/10.1007/978-3-030-66785-6_12

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-66785-6_12

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-66784-9

  • Online ISBN: 978-3-030-66785-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics