HITS-based attentional neural model for abstractive summarization

作者:

Highlights:

摘要

Automatic abstractive summary generation is still an open problem in natural language processing field. Conventional encoder–decoder model based abstractive summarization methods often suffer from repetition and semantic irrelevance. Recent studies apply traditional attention or graph-based attention on the encoder–decoder model to tackle the problem, under the assumption that all the sentences in the original document are indistinguishable from each other. But in a document, the same words in different sentences are not equally important, i.e., the words in a trivial sentence are less important than the words in a salient sentence. Based on it, we develop a HITS-based attention mechanism in this paper, which fully leverages sentence-level and word-level information by considering sentences and words in the original document as authorities and hubs. Based on it, we present a novel abstractive summarization method, with Kullback–Leibler (KL) divergence to refine the attention value, meanwhile we propose a comparison mechanism in summary generation to further improve the summarization performance. When evaluated on the CNN/Daily Mail and NYT datasets, the experimental results demonstrate the improvement of summarization performance and show the performance of our proposed method is comparable with that of the other summarization methods. Besides, we also conduct experiments on CORD-19 dataset (COVID-19 Open Research Dataset) which is a biomedical domain dataset, and the experimental results show great performance of our proposed method compared with that of the other state-of-the-art summarization methods.

论文关键词:HITS-based attention,Abstractive summarization,Comparison mechanism

论文评审过程:Received 18 August 2020, Revised 24 January 2021, Accepted 26 March 2021, Available online 27 March 2021, Version of Record 9 April 2021.

论文官网地址:https://doi.org/10.1016/j.knosys.2021.106996