Evaluation of Extractive and Abstract Methods in Text Summarization | SpringerLink
Skip to main content

Evaluation of Extractive and Abstract Methods in Text Summarization

  • Conference paper
  • First Online:
Data Science and Emerging Technologies (DaSET 2022)

Abstract

Text summarization has become very essential tool to record important points and has been used by several websites and applications to lessen length, difficulty, and to preserve the vital information of the original file. The requirement on well-organized and useful text summarization of the website content, news feed and other kinds of legal documents with judgments and predilection is the demand of the present requirement. Hence several attempts have been made to automate the summarizing process. The recent development and state of the art models in natural language processing demonstrated outstanding results in text summarization, however major focus of these analysis was on large dataset with large parameters. This study’s primary purpose is to evaluate the performance of ensemble abstractive and extractive models on text summarization. Combined core of BERT and PEGASUS models’ output were applied to LexRank model on News Summary dataset to evaluate the performance through ROUGE metric. The results showed the performance of combined and ensemble model is better than individual performance.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 26311
Price includes VAT (Japan)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 32889
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
JPY 32889
Price includes VAT (Japan)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)

    Google Scholar 

  2. See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. arXiv preprint arXiv:1704.04368 (2017)

  3. Aksenov, D., Julián, M., Peter, B., Robert, S., Leonhard, H., Georg, R.: Abstractive text summarization based on language model conditioning and locality modeling. arXiv arXiv:2003.13027 (2020)

  4. Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)

    Google Scholar 

  5. Fabbri, A.R., Kryściński, W., McCann, B., Xiong, C., Socher, R., Radev, D.: Summeval: re-evaluating summarization evaluation. Trans. Assoc. Comput. Linguist. 9, 391–409 (2021)

    Article  Google Scholar 

  6. Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems, vol. 27 (2014)

    Google Scholar 

  7. Cho, K., Van Merriënboer, B., Bahdanau, D., Bengio, Y.: On the properties of neural machine translation: encoder-decoder approaches. arXiv preprint arXiv:1409.1259 (2014)

  8. Agarwal, A., Lavie, A.: METEOR: an automatic metric for MT evaluation with high levels of correlation with human judgments. In: Proceedings of WMT-08 (2007)

    Google Scholar 

  9. Tjandra, A., Sakti, S., Nakamura, S.: Multi-scale alignment and contextual history for attention mechanism in sequence-to-sequence model. In: 2018 IEEE Spoken Language Technology Workshop (SLT), pp. 648–655. IEEE (2018)

    Google Scholar 

  10. Cohan, A., Goharian, N.: Revisiting summarization evaluation for scientific articles. arXiv preprint arXiv:1604.00400 (2016)

  11. Zhang, J., Zhao, Y., Saleh, M., Liu, P.: PEGASUS: pre-training with extracted gap-sentences for abstractive summarization. In: International Conference on Machine Learning, pp. 11328–11339. PMLR (2020)

    Google Scholar 

  12. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)

  13. Celikyilmaz, A., Bosselut, A., He, X., Choi, Y.: Deep communicating agents for abstractive summarization. arXiv preprint arXiv:1803.10357 (2018)

  14. Bouscarrat, L., Bonnefoy, A., Peel, T., Pereira, C.: STRASS: a light and effective method for extractive summarization based on sentence embeddings. arXiv preprint arXiv:1907.07323 (2019)

  15. Hao, Y., Dong, L., Wei, F., Xu, K.: Visualizing and understanding the effectiveness of BERT. arXiv preprint arXiv:1908.05620 (2019)

  16. Erkan, G., Radev, D.R.: LexRank: graph-based lexical centrality as salience in text summarization. J. Artif. Intell. Res. 22, 457–479 (2004)

    Article  Google Scholar 

  17. Lin, C.Y.: Rouge: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74–81 (2004)

    Google Scholar 

  18. Kaggle (2022). https://www.kaggle.com/

  19. Goodwin, T.R., Savery, M.E., Demner-Fushman, D.: Flight of the PEGASUS? Comparing transformers on few-shot and zero-shot multi-document abstractive summarization. In: Proceedings of COLING. International Conference on Computational Linguistics, vol. 2020, p. 5640. NIH Public Access (2020)

    Google Scholar 

  20. Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019)

  21. Liu, Y.: Fine-tune BERT for extractive summarization. arXiv preprint arXiv:1903.10318 (2019)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Dhiya Al-Jumeily OBE .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Lenka, R.K.B. et al. (2023). Evaluation of Extractive and Abstract Methods in Text Summarization. In: Wah, Y.B., Berry, M.W., Mohamed, A., Al-Jumeily, D. (eds) Data Science and Emerging Technologies. DaSET 2022. Lecture Notes on Data Engineering and Communications Technologies, vol 165. Springer, Singapore. https://doi.org/10.1007/978-981-99-0741-0_38

Download citation

Publish with us

Policies and ethics