Abstract
Obtaining valuable information from massive data efficiently has become our research goal in the era of Big Data. Text summarization technology has been continuously developed to meet this demand. Recent work has also shown that transformer-based pre-trained language models have achieved great success on various tasks in NLP. Aiming at the problem of Chinese news text summary generation and the application of Transformer structure on Chinese texts, this paper proposes a Chinese news headline generation model CNsum based on Transformer structure, and tests it on Chinese datasets such as THUCNews. The results of the conducted experiments show that CNsum achieves better ROUGE, BLEU and BERTScore scores than the baseline models, which verifies the outperformance of the model.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Allahyari, M., et al.: Text summarization techniques: a brief survey. Int. J. Adv. Comput. Sci. Appl. (IJACSA) 8, 397–405 (2017)
Iboi, H., Chua, S., Ranaivo-Malançon, B., Kulathuramaiyer, N.: Performance of opinion summarization towards extractive summarization. J. Telecommun. Electron. Comput. Eng. 9, 57–64 (2017)
Luhn, H.P.: The automatic creation of literature abstracts. IBM J. Res. Dev. 2(2), 157–165 (1958)
Syed, A., Gaol, F., Matsuo, T.: A survey of the state-of-the-art models in neural abstractive text summarization. IEEE Access 9, 13248–13265 (2021)
Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2, NIPS 2014, pp. 3104–3112. MIT Press, Cambridge (2014)
Cho, K., et al.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, pp. 1724–1734. Association for Computational Linguistics, October 2014. https://aclanthology.org/D14-1179
Vaswani, A., et al.: Attention is all you need. In: Guyon, I., et al. (eds.) Advances in Neural Information Processing Systems, vol. 30. Curran Associates Inc. (2017)
Li, J., Tang, T., Zhao, W.X., Wen, J.-R.: Pretrained language model for text generation: a survey. In: Zhou, Z.-H. (ed.) Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, International Joint Conferences on Artificial Intelligence Organization, pp. 4492–4499, Survey Track (2021)
Lin, C.-Y., Hovy, E.: Automatic evaluation of summaries using n-gram co-occurrence statistics. In: Proceedings of the 2003 Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pp. 150–157 (2003). https://aclanthology.org/N03-1020
Hu, B., Chen, Q., Zhu, F.: LCSTS: a large scale Chinese short text summarization dataset (2015)
Zhang, T., Kishore, V., Wu, F., Weinberger, K.Q., Artzi, Y.: BERTScore: evaluating text generation with BERT, arXiv, vol. abs/1904.09675 (2020)
Hua, L., Wan, X., Li, L.: Overview of the NLPCC 2017 shared task: single document summarization. In: Huang, X., Jiang, J., Zhao, D., Feng, Y., Hong, Yu. (eds.) NLPCC 2017. LNCS (LNAI), vol. 10619, pp. 942–947. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-73618-1_84
See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, Canada, pp. 1073–1083. Association for Computational Linguistics, July 2017. https://aclanthology.org/P17-1099
Hou, L.-W., Hu, P., Cao, W.-L.: Automatic Chinese abstractive summarization with topical keywords fusion. IEEE/CAA J. Automatica Sinica 45, 530–539 (2019)
Güneş, E., Radev, D.: Lexpagerank: prestige in multi-document text summarization. In: The 20th International Joint Conference on Artificial Intelligence, pp. 365–371 (2014)
Radev, D.R., et al.: MEAD - a platform for multidocument multilingual text summarization. In: LREC (2004)
Lin, H., Bilmes, J.: Multi-document summarization via budgeted maximization of submodular functions. In: Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, Los Angeles, California, pp. 912–920. Association for Computational Linguistics, June 2010
Chopra, S., Auli, M., Rush, A.M.: Abstractive sentence summarization with attentive recurrent neural networks. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, California, pp. 93–98. Association for Computational Linguistics, June 2016
Lin, C.-Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, Barcelona, Spain, pp. 74–81. Association for Computational Linguistics, July 2004
Papineni, K., Roukos, S., Ward, T., Zhu, W.-J.: Bleu: a method for automatic evaluation of machine translation. In: Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Philadelphia, Pennsylvania, USA, pp. 311–318. Association for Computational Linguistics, July 2002
Zhu, Q., Li, L., Bai, L., Hu, F.: Chinese text summarization based on fine-tuned GPT2. In: Mohiddin, M.K., Chen, S., EL-Zoghdy, S.F. (eds.) Third International Conference on Electronics and Communication; Network and Computer Technology (ECNCT 2021), SPIE 2022, vol. 12167, pp. 304–309. International Society for Optics and Photonics (2022). https://doi-org-s.libyc.nudt.edu.cn:443/10.1117/12.2629132
Kingma, D., Ba, J.: Adam: a method for stochastic optimization. Computer Science (2014)
Wolf, T., et al.: Huggingface’s transformers: state-of-the-art natural language processing, arXiv e-prints (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Zhao, Y., Huang, S., Zhou, D., Ding, Z., Wang, F., Nian, A. (2022). CNsum: Automatic Summarization for Chinese News Text. In: Wang, L., Segal, M., Chen, J., Qiu, T. (eds) Wireless Algorithms, Systems, and Applications. WASA 2022. Lecture Notes in Computer Science, vol 13472. Springer, Cham. https://doi.org/10.1007/978-3-031-19214-2_45
Download citation
DOI: https://doi.org/10.1007/978-3-031-19214-2_45
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-19213-5
Online ISBN: 978-3-031-19214-2
eBook Packages: Computer ScienceComputer Science (R0)