Skip to main content

Query Focused Multi-document Summarization Based on Five-Layered Graph and Universal Paraphrastic Embeddings

  • Conference paper
  • First Online:
Artificial Intelligence Trends in Intelligent Systems (CSOC 2017)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 573))

Included in the following conference series:

Abstract

Query focused multi-document summarization is a process of automatic query biased text compression of a document set. Lately, the graph-based and ranking methods have been intensively attracted the researchers from extractive document summarization domain. The uniform sentence connecteness or non-uniform document-sentence connecteness, such as sentence similarity weighted by document importance, were the main features used by work to date. Contrary, in this paper we present a novel five-layered heterogeneous graph model. It emphasizes not only sentence and document level relations but also the influence of lower level relations (e.g. a part of sentence similarity) and higher level relations (i.e. query to sentences similarity). Based on this model, we developed an iterative sentence ranking algorithm, based on the existing well known PageRank algorithm. Moreover, for text similarity calculations we used universal paraphrase embeddings that outperform various strong baselines on many text similarity tasks and many domains. Experiments are conducted on the DUC 2005 data sets and the ROUGE (Recall-Oriented Understudy for Gisting Evaluation) evaluation results demonstrate the advantages of the proposed approach.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 299.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 379.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    Document Understanding Conference (http://duc.nist.gov).

References

  1. Canhasi, E., Kononenko, I.: Weighted archetypal analysis of the multi-element graph for query-focused multi-document summarization. Expert Syst. Appl. 41(2), 535–543 (2014)

    Article  Google Scholar 

  2. Canhasi, E.: Fast document summarization using locality sensitive hashing and memory access efficient node ranking. Int. J. Electr. Comput. Eng. 6(3), 945 (2016)

    Google Scholar 

  3. Zwaan, R.A., Langston, M.C., Graesser, A.C.: The construction of situation models in narrative comprehension: an event-indexing model. Psychol. Sci. 6(5), 292–297 (1995)

    Article  Google Scholar 

  4. Wieting, J., Bansal, M., Gimpel, K., Livescu, K.: Towards universal paraphrastic sentence embeddings. arXiv preprint arXiv:1511.08198 (2015)

  5. Erkan, G., Radev, D.R.: Lexrank: graph-based lexical centrality as salience in text summarization. J. Artif. Intell. Res. (JAIR) 22, 457–479 (2004)

    Google Scholar 

  6. Mihalcea, R., Tarau, P.: Textrank: bringing order into text. In: EMNLP, pp. 404–411 (2004)

    Google Scholar 

  7. Brin, S., Page, L.: The anatomy of a large-scale hypertextual web search engine. Comput. Netw. 30(1–7), 107–117 (1998)

    Google Scholar 

  8. Kleinberg, J.M.: Authoritative sources in a hyperlinked environment. J. ACM 46(5), 604–632 (1999)

    Article  MathSciNet  MATH  Google Scholar 

  9. Carreras, X., Marque, L.: Introduction to the conll-2004 shared task: semantic role labeling. In: CoNLL, pp. 89–97 (2004)

    Google Scholar 

  10. Radev, D.R., Jing, H., Sty, M., Tam, D.: Centroid-based summarization of multiple documents. Inf. Process. Manage. 40(6), 919–938 (2004)

    Article  MATH  Google Scholar 

  11. Otterbacher, J., Erkan, G., Radev, D.R.: Biased lexrank: passage retrieval using random walks with question-based priors. Inf. Process. Manage. 45(1), 42–54 (2009)

    Article  Google Scholar 

  12. Wei, F., Li, W., Qin, L., He, Y.: A document-sensitive graph model for multi-document summarization. Knowl. Inf. Syst. 22(2), 245–259 (2010)

    Article  Google Scholar 

  13. Wan, X.: Document-based HITS model for multi-document summarization. In: Ho, T.-B., Zhou, Z.-H. (eds.) PRICAI 2008. LNCS (LNAI), vol. 5351, pp. 454–465. Springer, Heidelberg (2008). doi:10.1007/978-3-540-89197-0_42

    Chapter  Google Scholar 

  14. Lin, C.-Y., Hovy, E.H.: Automatic evaluation of summaries using n-gram co-occurrence statistics. In: HLT-NAACL (2003)

    Google Scholar 

  15. Canhasi, E., Kononenko, I.: Weighted hierarchical archetypal analysis for multi-document summarization. Comput. Speech Lang. 37, 24–46 (2016)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ercan Canhasi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Canhasi, E. (2017). Query Focused Multi-document Summarization Based on Five-Layered Graph and Universal Paraphrastic Embeddings. In: Silhavy, R., Senkerik, R., Kominkova Oplatkova, Z., Prokopova, Z., Silhavy, P. (eds) Artificial Intelligence Trends in Intelligent Systems. CSOC 2017. Advances in Intelligent Systems and Computing, vol 573. Springer, Cham. https://doi.org/10.1007/978-3-319-57261-1_22

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-57261-1_22

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-57260-4

  • Online ISBN: 978-3-319-57261-1

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics