Unsupervised extractive multi-document summarization method based on transfer learning from BERT multi-task fine-tuning

被引:26
作者
Lamsiyah, Salima [1 ]
El Mahdaouy, Abdelkader [3 ]
Ouatik, Said El Alaoui [1 ,2 ]
Espinasse, Bernard [4 ]
机构
[1] Sidi Mohamed Ben Abdellah Univ, FSDM, Lab Informat Signals Automat & Cognitivism, BP 1796, Fez Atlas 30003, Morocco
[2] Ibn Tofail Univ, Natl Sch Appl Sci, Lab Engn Sci, Kenitra, Morocco
[3] Mohammed VI Polytech Univ UM6P, Sch Comp Sci UM6P CS, Ben Guerir, Morocco
[4] Univ Toulon & Var, Aix Marseille Univ, CNRS, LIS,UMR 7020, Toulon, France
关键词
BERT fine-tuning; multi-document summarization; multi-task learning; sentence representation learning; transfer learning; SENTENCE SCORING TECHNIQUES;
D O I
10.1177/0165551521990616
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Text representation is a fundamental cornerstone that impacts the effectiveness of several text summarization methods. Transfer learning using pre-trained word embedding models has shown promising results. However, most of these representations do not consider the order and the semantic relationships between words in a sentence, and thus they do not carry the meaning of a full sentence. To overcome this issue, the current study proposes an unsupervised method for extractive multi-document summarization based on transfer learning from BERT sentence embedding model. Moreover, to improve sentence representation learning, we fine-tune BERT model on supervised intermediate tasks from GLUE benchmark datasets using single-task and multi-task fine-tuning methods. Experiments are performed on the standard DUC'2002-2004 datasets. The obtained results show that our method has significantly outperformed several baseline methods and achieves a comparable and sometimes better performance than the recent state-of-the-art deep learning-based methods. Furthermore, the results show that fine-tuning BERT using multi-task learning has considerably improved the performance.
引用
收藏
页码:164 / 182
页数:19
相关论文
共 74 条
[1]  
Bouscarrat L, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019:): STUDENT RESEARCH WORKSHOP, P243
[2]  
Brockett C., P 3 INT WORKSH PAR I
[3]  
Cao ZQ, 2015, AAAI CONF ARTIF INTE, P2153
[4]  
Cao ZQ, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL) AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (IJCNLP), VOL 2, P829
[5]   Multitask learning [J].
Caruana, R .
MACHINE LEARNING, 1997, 28 (01) :41-75
[6]  
Cer Daniel., 2017, SEMEVAL ACL, P1, DOI [10.18653/v1/S17-2001, DOI 10.18653/V1/S17-2001]
[7]  
Conneau Alexis., 2017, Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, P670, DOI [10.18653/v1/D17-1070, 10.18653/v1/d17-1070, DOI 10.18653/V1/D17-1070]
[8]  
Conroy JM., 2004, P DOC UND C DUC 2004
[9]   OCCAMS - An Optimal Combinatorial Covering Algorithm for Multi-document Summarization [J].
Davis, Sashka T. ;
Conroy, John M. ;
Schlesinger, Judith D. .
12TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW 2012), 2012, :454-463
[10]  
Denil M., 2014, Extraction of salient sentences from labelled documents