Hybrid multi-document summarization using pre-trained language models

被引:19
作者
Ghadimi, Alireza [1 ]
Beigy, Hamid [2 ]
机构
[1] Acad Ctr Educ Culture & Res ACECR, Res Inst Informat & Commun Technol, Tehran, Iran
[2] Sharif Univ Technol, Dept Comp Engn, Tehran, Iran
关键词
Pre-trained language models; Extractive summarization; Abstractive summarization; Determinantal point process; Deep submodular network; SENTENCE FUSION;
D O I
10.1016/j.eswa.2021.116292
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
multi-document summarization is a type of automatic text summarization. It obtains information from multiple documents and generates a human-like summary from them. In this paper, we propose an abstractive multi-document summarization method called HMSumm. The proposed method is a combination of extractive and abstractive summarization approaches. First, it constructs an extractive summary from multiple input documents, and then uses it to generate the abstractive summary. Redundant information, which is a global problem in multi-document summarization, is managed in the first step. Specifically, the determinantal point process (DPP) is used to deal with redundancy. This step also controls the length of the input sequence for the abstractive summarization process. This step has two effects. The first is to reduce the computational time. The second is to preserve the important parts of the input documents for an abstractive summarizer. We employ a deep submodular network (DSN) to determine the quality of the sentences in the extractive summary, and use BERT-based similarities to compute the redundancy. The obtained extractive summary is fed into BART and T5 pre-trained models to generate two abstractive summaries. We use the diversity of sentences in each summary to select one of them as the final abstractive summary. To evaluate the performance of HMSumm, we use both human evaluations and ROUGE-based assessments, and compare it with several state-of-the-art methods. We use DUC 2002, DUC 2004, Multi-News, and CNN/DailyMail datasets to evaluate the algorithms. The experimental results show that HMSumm outperforms the related state-of-the-art algorithms.
引用
收藏
页数:10
相关论文
共 63 条
[1]  
[Anonymous], 2017, P 2017 C EMP METH NA
[2]   Sentence fusion for multidocument news summarization [J].
Barzilay, R ;
McKeown, KR .
COMPUTATIONAL LINGUISTICS, 2005, 31 (03) :297-327
[3]   A neural probabilistic language model [J].
Bengio, Y ;
Ducharme, R ;
Vincent, P ;
Jauvin, C .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (06) :1137-1155
[4]   Eynard-Mehta theorem, schur process, and their pfaffian analogs [J].
Borodin, A ;
Rains, EM .
JOURNAL OF STATISTICAL PHYSICS, 2005, 121 (3-4) :291-317
[5]   A LIMITED MEMORY ALGORITHM FOR BOUND CONSTRAINED OPTIMIZATION [J].
BYRD, RH ;
LU, PH ;
NOCEDAL, J ;
ZHU, CY .
SIAM JOURNAL ON SCIENTIFIC COMPUTING, 1995, 16 (05) :1190-1208
[6]  
Chen Qian, 2016, P 25 INT JOINT C ART, P2754
[7]  
Cheng JP, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P484
[8]  
Cho S, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P1027
[9]  
Chopra S., 2016, P 2016 ACL, P93, DOI 10.18653/v1/n16-1012
[10]   Sentence Compression as Tree Transduction [J].
Cohn, Trevor ;
Lapata, Mirella .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2009, 34 :637-674