Mixed-Lingual Pre-training for Cross-lingual Summarization

被引:0
作者
Xu, Ruochen [1 ]
Zhu, Chenguang [1 ]
Shi, Yu [1 ]
Zeng, Michael [1 ]
Huang, Xuedong [1 ]
机构
[1] Microsoft Cognit Serv Res Grp, Redmond, WA 98052 USA
来源
1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020) | 2020年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-lingual Summarization (CLS) aims at producing a summary in the target language for an article in the source language. Traditional solutions employ a twostep approach, i.e. translate!summarize or summarize!translate. Recently, end-to-end models have achieved better results, but these approaches are mostly limited by their dependence on large-scale labeled data. We propose a solution based on mixed-lingual pretraining that leverages both cross-lingual tasks such as translation and monolingual tasks like masked language models. Thus, our model can leverage the massive monolingual data to enhance its modeling of language. Moreover, the architecture has no task-specific components, which saves memory and increases optimization efficiency. We show in experiments that this pre-training scheme can effectively boost the performance of cross-lingual summarization. In Neural Cross-Lingual Summarization (NCLS) (Zhu et al., 2019b) dataset, our model achieves an improvement of 2.82 (English to Chinese) and 1.15 (Chinese to English) ROUGE-1 scores over state-of-the-art results.
引用
收藏
页码:536 / 541
页数:6
相关论文
共 50 条
[21]   Product-oriented Machine Translation with Cross-modal Cross-lingual Pre-training [J].
Song, Yuqing ;
Chen, Shizhe ;
Jin, Qin ;
Luo, Wei ;
Xie, Jun ;
Huang, Fei .
PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, :2843-2852
[22]   XLM-E: Cross-lingual Language Model Pre-training via ELECTRA [J].
Chi, Zewen ;
Huang, Shaohan ;
Dong, Li ;
Ma, Shuming ;
Zheng, Bo ;
Singhal, Saksham ;
Bajaj, Payal ;
Song, Xia ;
Mao, Xian-Ling ;
Huang, Heyan ;
Wei, Furu .
PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, :6170-6182
[23]   INFOXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training [J].
Chi, Zewen ;
Dong, Li ;
Wei, Furu ;
Yang, Nan ;
Singhal, Saksham ;
Wang, Wenhui ;
Song, Xia ;
Mao, Xian-Ling ;
Huang, Heyan ;
Zhou, Ming .
2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, :3576-3588
[24]   PTEKC: pre-training with event knowledge of ConceptNet for cross-lingual event causality identification [J].
Zhu, Enchang ;
Yu, Zhengtao ;
Huang, Yuxin ;
Gao, Shengxiang ;
Xian, Yantuan .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2025, 16 (03) :1859-1872
[25]   A Cross-Lingual Summarization method based on cross-lingual Fact-relationship Graph Generation [J].
Zhang, Yongbing ;
Gao, Shengxiang ;
Huang, Yuxin ;
Tan, Kaiwen ;
Yu, Zhengtao .
PATTERN RECOGNITION, 2024, 146
[26]   Cross-Lingual Speech-to-Text Summarization [J].
Pontes, Elvys Linhares ;
Gonzalez-Gallardo, Carlos-Emiliano ;
Torres-Moreno, Juan-Manuel ;
Huet, Stephane .
MULTIMEDIA AND NETWORK INFORMATION SYSTEMS, 2019, 833 :385-395
[27]   Cross-lingual extreme summarization of scholarly documents [J].
Takeshita, Sotaro ;
Green, Tommaso ;
Friedrich, Niklas ;
Eckert, Kai ;
Ponzetto, Simone Paolo .
INTERNATIONAL JOURNAL ON DIGITAL LIBRARIES, 2024, 25 (02) :249-271
[28]   A Robust Abstractive System for Cross-Lingual Summarization [J].
Ouyang, Jessica ;
Song, Boya ;
McKeown, Kathleen .
2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, :2025-2031
[29]   Few-Shot Cross-Lingual Stance Detection with Sentiment-Based Pre-training [J].
Hardalov, Momchil ;
Arora, Arnav ;
Nakov, Preslav ;
Augenstein, Isabelle .
THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, :10729-10737
[30]   XLM-K: Improving Cross-Lingual Language Model Pre-training with Multilingual Knowledge [J].
Jiang, Xiaoze ;
Liang, Yaobo ;
Chen, Weizhu ;
Duan, Nan .
THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, :10840-10848