Abstract
Two of the most crucial issues for text summarization to generate faithful summaries are to make use of knowledge beyond text and to make use of cross-sentence relations in text. Intuitive ways for the two issues are knowledge graph (KG) and graph neural network (GNN), respectively. Entities are semantic units in text and in KG. This paper focuses on both issues by leveraging entities mentioned in text to connect GNN and KG for summarization. Firstly, entities are leveraged to construct a sentence-entity graph with weighted multi-type edges to model sentence relations, and a relational heterogeneous GNN for summarization is proposed to calculate node encodings. Secondly, entities are leveraged to link the graph to KG to collect knowledge. Thirdly, entities guide a two-step summarization framework defining a multitask selector to select salient sentences and entities, and using an entity-focused abstractor to compress the sentences. GNN is connected with KG by constructing sentence-entity graphs where entity–entity edges are built based on KG, initializing entity embeddings on KG, and training entity embeddings using entity–entity edges. The relational heterogeneous GNN utilizes both edge weights and edge types in GNN to calculate graphs with weighted multi-type edges. Experiments show the proposed method outperforms extractive baselines including the HGNN-based HGNNSum and abstractive baselines including the entity-driven SENECA on CNN/DM, and outperforms most baselines on NYT50. Experiments on sub-datasets show the density of sentence-entity edges greatly influences the performance of the proposed method. The greater the density, the better the performance. Ablations show effectiveness of the method.
Similar content being viewed by others
Data availability
The datasets and codes are available at https://github.com/jingqiangchen/kbsumm.
References
Zhuge H (2016) Multi-dimensional summarization in cyber-physical society. Morgan Kaufmann
Wang D, Liu P, Zheng Y, et al. (2020) Heterogeneous graph neural networks for extractive document summarization. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 6209–6219.
Sharma E, Huang L, Hu Z, et al. (2019) An entity-driven framework for abstractive summarization. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 3271–3282.
Cheng J, Lapata M (2016) Neural summarization by extracting sentences and words. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 484–494.
Nallapati R, Zhai F, Zhou B (2017) Summarunner: a recurrent neural network based sequence model for extractive summarization of documents. In: Proceedings of the AAAI Conference on Artificial Intelligence, 3075–3081.
Liu Y, Lapata M (2019) Hierarchical transformers for multi-document summarization. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 5070–5081.
Li H, Zhu J, Zhang J, et al. (2018) Ensure the correctness of the summary: incorporate entailment knowledge into abstractive sentence summarization. In: Proceedings of the 27th International Conference on Computational Linguistics. 2018: 1430–1441.
Feng X, Feng X, Qin B. (2021) Incorporating commonsense knowledge into abstractive dialogue summarization via heterogeneous graph networks. In: China National Conference on Chinese Computational Linguistics. Springer, Cham, 2021: 127–142.
Li W, Xu J, He Y, et al (2019) Coherent comments generation for chinese articles with a graph-to-sequence model. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 4843–4852.
Gunel B, Zhu C, Zeng M, et al (2020) Mind The Facts: knowledge-boosted coherent abstractive Text summarization. arXiv preprint arXiv:2006.15435.
Zhang Z, Han X, Liu Z, et al (2019) ERNIE: Enhanced language representation with informative entities. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 1441–1451.
Liu W, Zhou P, Zhao Z, et al (2020) K-bert: enabling language representation with knowledge graph. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 2901–2908.
Erkan G, Radev DR (2004) Lexrank: graph-based lexical centrality as salience in text summarization. J ARTIF INTELL RES 22:457–479
Mihalcea R, Tarau P (2004) Textrank: bringing order into text. In: Proceedings of the 2004 conference on empirical methods in natural language processing, 404–411.
Kipf TN, Welling M (2016) Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907.
Zhao L, Xu W, Guo J. (2020) Improving abstractive dialogue summarization with graph structures and topic words. In: Proceedings of the 28th International Conference on Computational Linguistics. 2020: 437–449.
Yasunaga M, Zhang R, Meelu K, et al (2017) Graph-based Neural Multi-Document Summarization. In: Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL), 452–462.
Xu J, Gan Z, Cheng Y, et al (2020) Discourse-Aware Neural Extractive Text Summarization. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 5021–5031.
Schlichtkrull M, Kipf TN, Bloem P et al (2018) Modeling relational data with graph convolutional networks. European semantic web conference. Springer, Cham, pp 593–607
Hoffart J, Yosef MA, Bordino I, et al (2011) Robust disambiguation of named entities in text. In: Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, 782–792.
Nguyen DB, Hoffart J, Theobald M et al (2014) AIDA-light: high-throughput named-entity disambiguation. Workshop on Linked Data on the Web 2014:1–10
Cao Y, Hou L, Li J, et al (2018) Neural collective entity linking. In: Proceedings of the 27th International Conference on Computational Linguistics, 675–686.
Ristoski P, Rosati J, Di Noia T et al (2019) RDF2Vec: RDF graph embeddings and their applications. SEMANT WEB 10(4):721–752
Nenkova A (2008) Entity-driven rewrite for multi-document summarization. In: Proceedings of the Third International Joint Conference on Natural Language Processing, 118–125.
Inan E (2021) Somun: entity-centric summarization incorporating pre-trained language models. NEURAL COMPUT APPL 33:5301–5311. https://doi.org/10.1007/s00521-020-05319-2
Parveen D, Ramsl HM, Strube M (2015) Topical coherence for graph-based extractive summarization. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 1949–1954.
Huang L, Wu L, Wang L. (2020) Knowledge Graph-Augmented Abstractive Summarization with Semantic-Driven Cloze Reward. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 2020: 5094–5107.
Zhuge H (2020) Cyber-physical-social intelligence on human-machine-nature symbiosis. Springer
Bordes A, Usunier N, Garcia-Durán A, et al (2013) Translating embeddings for modeling multi-relational data. In: Proceedings of the 26th International Conference on Neural Information Processing, 2787–2795.
Vaswani A, Shazeer N, Parmar N, et al (2017) Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, 6000–6010.
Devlin J, Chang MW, Lee K, et al (2019) BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 4171–4186.
Veličković P, Cucurull G, Casanova A, et al (2017) Graph attention networks, arXiv preprint arXiv:1710.10903.
Gilmer J, Schoenholz SS, Riley PF, et al (2017) Neural message passing for quantum chemistry. In: International Conference on Machine Learning, PMLR, 1263–1272.
Fernandes P, Allamanis M, Brockschmidt M (2018) Structured neural summarization. arXiv preprint arXiv:1811.01824.
Zhang C, Song D, Huang C, et al (2019) Heterogeneous graph neural network. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 793–803.
Shi J, Ji H, Shi C, et al (2020) Heterogeneous Graph Neural Network for Recommendation. arXiv preprint arXiv:2009.00799.
Wang S, Chen Z, Li D, et al (2019) Attentional heterogeneous graph neural network: application to program reidentification. In: Proceedings of the 2019 SIAM International Conference on Data Mining. Society for Industrial and Applied Mathematics, 693–701.
Hamilton WL, Ying R, Leskovec J (2017) Inductive representation learning on large graphs. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, 1025–1035.
Wang, X, Ji H, Shi C, Wang B, Ye Y, Cui P, Yu PS (2019) Heterogeneous graph attention network. In: The World Wide Web Conference, 2022–2032.
Kang S (2021) Product failure prediction with missing data using graph neural networks. NEURAL COMPUT APPL 33:7225–7234. https://doi.org/10.1007/s00521-020-05486-2
CSIRO’s Data61 (2018) Stellargraph machine learning library. https://github.com/stellargraph/stellargraph
Xiao W, Carenini G (2019) Extractive summarization of long documents by combining global and local context. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 3002–3012.
Zhong M, Liu P, Wang D, et al (2019) Searching for effective neural extractive summarization: what works and what’s next. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 1049–1058.
Li S, Xu J (2021) A two-step abstractive summarization model with asynchronous and enriched-information decoding. NEURAL COMPUT APPL 33:1159–1170. https://doi.org/10.1007/s00521-020-05005-3
Chen YC, Bansal M (2018) Fast abstractive summarization with reinforce-selected sentence rewriting. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics 1: 675–686.
Li C, Xu W, Li S, et al (2018) Guiding generation for abstractive text summarization based on key information guide network. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2: 55–60.
Gehrmann S, Deng Y, Rush AM (2018) Bottom-up abstractive summarization. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 4098–4109.
Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473.
See A, Liu PJ, Manning CD (2017) Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics 1: 1073–1083.
Hoffart J, Suchanek FM, Berberich K et al (2013) YAGO2: a spatially and temporally enhanced knowledge base from Wikipedia. ARTIF INTELL 194:28–61. https://doi.org/10.1016/j.artint.2012.06.001
Manning CD, Surdeanu M, Bauer J, et al (2014) The Stanford CoreNLP natural language processing toolkit. In: Proceedings of 52nd annual meeting of the association for computational linguistics: system demonstrations, 55–60.
Rennie SJ, Marcheret E, Mroueh Y, et al (2017) Self-critical sequence training for image captioning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 7008–7024.
Lin CY (2004) Rouge: a package for automatic evaluation of summaries. Text summarization branches out, 74–81.
Hermann KM, Kočiský T, Grefenstette E, et al (2015) Teaching machines to read and comprehend. In: Proceedings of the 28th International Conference on Neural Information Processing Systems-1: 1693–1701.
Durrett G, Berg-Kirkpatrick T, Klein D (2016) Learning-Based Single-Document Summarization with Compression and Anaphoricity Constraints. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 1: 1998–2008.
Sandhaus E (2008) The new york times annotated corpus. Linguistic Data Consortium, Philadelphia
Mikolov T, Sutskever I, Chen K, et al (2013) Distributed representations of words and phrases and their compositionality. arXiv preprint arXiv:1310.4546.
Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980.
Liu Y, Lapata M (2019) Text Summarization with Pretrained Encoders. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 3721–3731.
Xu J, Durrett G (2019) Neural Extractive Text Summarization with Syntactic Compression. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 3283–3294.
Luo L, Ao X, Song Y, et al (2019) Reading like HER: Human reading inspired extractive summarization. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 3024–3034.
Paulus R, Xiong C, Socher R (2017) A deep reinforced model for abstractive summarization. arXiv preprint arXiv:1705.04304.
Liu Y, Titov I, Lapata M (2019) Single document summarization as tree induction. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 1: 1745–1755.
Acknowledgements
This research was sponsored by the National Natural Science Foundation of China (No.61806101).
Funding
The Funding was provided by National Natural Science Foundation of China, (61806101), Jingqiang Chen
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The author(s) of this manuscript have no conflict of interest, financial, or otherwise. And there are no other relationships or activities that could appear to have influenced the submitted work.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Chen, J. An entity-guided text summarization framework with relational heterogeneous graph neural network. Neural Comput & Applic 36, 3613–3630 (2024). https://doi.org/10.1007/s00521-023-09247-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-023-09247-9