Abstract
Effective unsupervised query-focused extractive summarization systems use query-specific features along with short-range language models (LMs) in sentence ranking and selection summarization subtasks. We hypothesize that applying long-span n-gram-based and neural LMs that better capture larger context can help improve these subtasks. Hence, we outline the first attempt to apply long-span models to a query-focused summarization task in an unsupervised setting. We also propose the A cross S entence B oundary LSTM-based LMs, ASB LSTM and bi ASB LSTM, that is geared towards the query-focused summarization subtasks. Intrinsic and extrinsic experiments on a real word corpus with 100 Wikipedia event descriptions as queries show that using the long-span models applied in an integer linear programming (ILP) formulation of MMR criterion are the most effective against several state-of-the-art baseline methods from the literature.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Cao, Z., et al.: AttSum: joint learning of focusing and summarization with neural attention. arXiv preprint arXiv:1604.00125 (2016)
Nallapati, R., et al.: Classify or select: neural architectures for extractive document summarization. arXiv preprint arXiv:1611.04244 (2016)
Mishra, A., et al.: Event digest: a holistic view on past events. In: SIGIR (2016)
Zhong, S., et al.: Query-oriented unsupervised multi-document summarization via deep learning model. Expert Syst. Appl. 42(21) (2015)
Yousefi-Azar, M., et al.: Text summarization using unsupervised deep learning. Expert Syst. Appl. 68, 93–105 (2017)
Cao, Z., et al.: Ranking with recursive neural networks and its application to multi-document summarization. In: AAAI (2015)
Palangi, H., et al.: Deep sentence embedding using long short-term memory networks: analysis and application to information retrieval. In: TASLP (2016)
Momtazi, S., et al.: Trained trigger language model for sentence retrieval in QA: bridging the vocabulary gap. In: CIKM (2011)
McDonald, R.: A study of global inference algorithms in multi-document summarization. In: Amati, G., Carpineto, C., Romano, G. (eds.) ECIR 2007. LNCS, vol. 4425, pp. 557–564. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-71496-5_51
Gillick, D., et al.: A scalable global model for summarization. In: ILP-NAACL-HLT (2009)
Riedhammer, K., et al.: Long story short - global unsupervised models for keyphrase based meeting summarization. Speech Commun. 52(10), 801–815 (2010)
Carbonell, J., et al.: The use of MMR, diversity-based reranking for reordering documents and producing summaries. In: SIGIR (1998)
Oualil, Y., et al.: Long-short range context neural networks for language modeling. arXiv preprint arXiv:1708.06555 (2017)
English Gigaword Corpus. https://catalog.ldc.upenn.edu/ldc2003t05
Zhai, C.X., et al.: Statistical language models for information retrieval. Synth. Lect. Hum. Lang. Technol. 1(1), 1–141 (2008)
Radev, D.R., et al.: MEAD-a platform for multidocument multilingual text summarization. In: LREC (2004)
Riedhammer, K., et al.: Packing the meeting summarization knapsack. In: INTERSPEECH (2008)
Guo, J., et al.: A deep relevance matching model for ad-hoc retrieval. In: CIKM (2016)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Singh, M., Mishra, A., Oualil, Y., Berberich, K., Klakow, D. (2018). Long-Span Language Models for Query-Focused Unsupervised Extractive Text Summarization. In: Pasi, G., Piwowarski, B., Azzopardi, L., Hanbury, A. (eds) Advances in Information Retrieval. ECIR 2018. Lecture Notes in Computer Science(), vol 10772. Springer, Cham. https://doi.org/10.1007/978-3-319-76941-7_59
Download citation
DOI: https://doi.org/10.1007/978-3-319-76941-7_59
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-76940-0
Online ISBN: 978-3-319-76941-7
eBook Packages: Computer ScienceComputer Science (R0)