Skip to main content

Highlighted Word Encoding for Abstractive Text Summarization

  • Conference paper
  • First Online:
Intelligent Human Computer Interaction (IHCI 2019)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 11886))

Included in the following conference series:

Abstract

The proposed model unites the robustness of the extractive and abstractive summarization strategies. Three tasks indispensable to automatic summarization, namely, apprehension, extraction, and abstraction, are performed by two specially designed networks, the highlighter RNN and the generator RNN. While the highlighter RNN collectively performs the task of highlighting and extraction for identifying the salient facts in the input text, the generator RNN fabricates the summary based on those facts. The summary is generated using word-level extraction with the help of term-frequency inverse document frequency (TFIDF) ranking factor. The union of the two strategies proves to surpass the ROUGE score results on the Gigaword dataset as compared to the simple abstractive approach for summarization.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. CoRR, abs/1409.0473 (2014)

    Google Scholar 

  2. Saggion, H., Poibeau, T.: Automatic Text Summarization: Past Present and Future, pp. 3–13. Springer, Heidelberg (2016)

    Google Scholar 

  3. Barrios, F., Lopez, F., Argerich, L., Wachenchauzer, R.: Variations of the Similarity Function of TextRank for Automated Summarization. arXiv preprint arXiv:1602.03606v1 (2016)

  4. Durrett, G., Berg-Kirkpatrick, T., Klein, D.: Learning-Based Single-Document Summarization with Compression and Anaphoricity Constraints. arXiv preprint arXiv:1603.08887v2 (2016)

  5. Nallapati, R., Zhai, F., Zhou, B.: Summarunner: a recurrent neural network-based sequence model for extractive summarization of documents (2017)

    Google Scholar 

  6. Cheng, J., Lapata, M.: Neural summarization by extracting sentences and words. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), vol. 1, pp. 484–494 (2016)

    Google Scholar 

  7. Nallapati, R., Zhou, B., Ma, M.: Classify or select: Neural architectures for extractive document summarization. arXiv preprint arXiv:1611.04244 (2016a)

  8. Narayan, S., Papasarantopoulos, N., Lapata, M., Cohen, S.B.: Neural extractive summarization with side information. arXiv preprint arXiv:1704.04530 (2017)

  9. Nallapati, R., Zhou, B., dos Santos, C., Gulcehre, C., Xiang, B.: Abstractive text summarization using sequence-to-sequence RNNs and beyond. In Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning, pp. 280–290 (2016b)

    Google Scholar 

  10. Paulus, R., Xiong, C., Socher, R.: A deep reinforced model for abstractive summarization. In: Proceedings of the 2018 International Conference on Learning Representations (2017)

    Google Scholar 

  11. Rush, A.M., Chopra, S., Weston, J.: A neural attention model for abstractive sentence summarization. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 379–389 (2015)

    Google Scholar 

  12. Fan, A., Grangier, D., Auli, M.: Controllable abstractive summarization. arXiv preprint arXiv:1711.05217 (2017)

  13. Zhou, Q., Yang, N., Wei, F., Zhou, M.: Selective Encoding for Abstractive Sentence Summarization. arXiv preprint arXiv:1704.07073v1 (2017)

  14. Lin, C.-Y.: ROUGE: a package for automatic evaluation of summaries. In: Proceedings of the ACL Workshop: Text Summarization Branches Out (2004)

    Google Scholar 

  15. Chopra, S., Auli, M., Rush, A.M.: Abstractive sentence summarization with attentive recurrent neural networks. In: Proceedings of the ACL Workshop: Text Summarization Branches Out (2016)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Daisy Monika Lal .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Lal, D.M., Singh, K.P., Tiwary, U.S. (2020). Highlighted Word Encoding for Abstractive Text Summarization. In: Tiwary, U., Chaudhury, S. (eds) Intelligent Human Computer Interaction. IHCI 2019. Lecture Notes in Computer Science(), vol 11886. Springer, Cham. https://doi.org/10.1007/978-3-030-44689-5_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-44689-5_7

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-44688-8

  • Online ISBN: 978-3-030-44689-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics