Query focused summarization via relevance distillation

Ye Yue, Yuanli Li, Jia ao Zhan, Yang Gao*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)
Plum Print visual indicator of research metrics
  • Citations
    • Citation Indexes: 1
  • Captures
    • Readers: 3
see details

Abstract

Creating a short version of a concise and relevant summary regarding a specific query can broadly meet a user’s information needs in many areas. In a summarization system, the extractive technique is attractive because it is simple and fast and produces reliable outputs. Salience and relevance are two key points for the extractive summarization. The majority of existing approaches to achieving them are augmenting input features, incorporating additional attention, or expanding the training scales. Yet, there is much unsupervised but query-related knowledge needs better exploration. To this end, in this paper, we frame the query-focused document summarization as a combination of salience prediction and relevance prediction. Concretely, in addition to the oracle summary set for the salience task, we further create a pseudo-summary set regarding user-specific queries (i.e., title or image captions as the query) for the relevance task. Then, based on a modified BERT fine-tune summarization, we propose two methods, called guidance and distillation, respectively. Specifically, the guidance training essentially shares salient information to reinforce the useful contextual representations in a two-stage training with the salience-and-relevance objective. For the distillation, we propose a new “guide-student” learning paradigm that the relevance knowledge of the query is distilled and transferred from a guide model to a salience-oriented student model. Experiment results demonstrate that guidance training prevails at improving the salience of the summary and distillation training is significantly advanced at relevance learning. Both of them achieve the best state of the arts in unsupervised query-focused settings of CNN and DailyMail dataset.

Original languageEnglish
Pages (from-to)16543-16557
Number of pages15
JournalNeural Computing and Applications
Volume35
Issue number22
DOIs
Publication statusPublished - Aug 2023

Keywords

  • Document summarization
  • Knowledge distillation
  • Unsupervised method

Fingerprint

Dive into the research topics of 'Query focused summarization via relevance distillation'. Together they form a unique fingerprint.

Cite this

Yue, Y., Li, Y., Zhan, J. A., & Gao, Y. (2023). Query focused summarization via relevance distillation. Neural Computing and Applications, 35(22), 16543-16557. https://doi.org/10.1007/s00521-023-08525-w