AdaptSum: Towards Low-Resource Domain Adaptation for Abstractive Summarization

被引:0
|
作者
Yu, Tiezheng [1 ]
Liu, Zihan [1 ]
Fung, Pascale [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Ctr Artificial Intelligence Res CAiRE, Dept Elect & Comp Engn, Clear Water Bay, Hong Kong, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art abstractive summarization models generally rely on extensive labeled data, which lowers their generalization ability on domains where such data are not available. In this paper, we present a study of domain adaptation for the abstractive summarization task across six diverse target domains in a low-resource setting. Specifically, we investigate the second phase of pre-training on large-scale generative models under three different settings: 1) source domain pre-training; 2) domain-adaptive pre-training; and 3) task-adaptive pre-training Experiments show that the effectiveness of pre-training is correlated with the similarity between the pre-training data and the target domain task. Moreover, we find that continuing pre-training could lead to the pre-trained model's catastrophic forgetting, and a learning method with less forgetting can alleviate this issue. Furthermore, results illustrate that a huge gap still exists between the low-resource and high-resource settings, which highlights the need for more advanced domain adaptation methods for the abstractive summarization task.
引用
收藏
页码:5892 / 5904
页数:13
相关论文
共 50 条
  • [1] Exploring Multitask Learning for Low-Resource Abstractive Summarization
    Magooda, Ahmed
    Elaraby, Mohamed
    Litman, Diane
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 1652 - 1661
  • [2] SPEC: Summary Preference Decomposition for Low-Resource Abstractive Summarization
    Chen, Yi-Syuan
    Song, Yun-Zhu
    Shuai, Hong-Han
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 603 - 618
  • [3] Meta-Transfer Learning for Low-Resource Abstractive Summarization
    Chen, Yi-Syuan
    Shuai, Hong-Han
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 12692 - 12700
  • [4] Lightweight Meta-Learning for Low-Resource Abstractive Summarization
    Huh, Taehun
    Ko, Youngjoong
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2629 - 2633
  • [5] Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization
    Chernyshev, Daniil
    Dobrov, Boris
    IEEE ACCESS, 2024, 12 : 47219 - 47230
  • [6] Semantic Self-Segmentation for Abstractive Summarization of Long Documents in Low-Resource Regimes
    Moro, Gianluca
    Ragazzi, Luca
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11085 - 11093
  • [7] Towards Two-Step Fine-Tuned Abstractive Summarization for Low-Resource Language Using Transformer T5
    Nasution, Salhazan
    Ferdiana, Ridi
    Hartanto, Rudy
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2025, 16 (02) : 1220 - 1230
  • [8] Efficient framework for low-resource abstractive summarization by meta-transfer learning and pointer-generator networks
    Huh, Taehun
    Ko, Youngjoong
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 234
  • [9] Low-Resource Adaptation of Open-Domain Generative Chatbots
    Gerhard-Young, Greyson
    Anantha, Raviteja
    Chappidi, Srinivas
    Hoffmeister, Bjorn
    PROCEEDINGS OF THE SECOND DIALDOC WORKSHOP ON DOCUMENT-GROUNDED DIALOGUE AND CONVERSATIONAL QUESTION ANSWERING (DIALDOC 2022), 2022, : 23 - 30
  • [10] Low-Resource Dialogue Summarization with Domain-Agnostic Multi-Source Pretraining
    Zou, Yicheng
    Zhu, Bolin
    Hu, Xingwu
    Gui, Tao
    Zhang, Qi
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 80 - 91