Balanced Meta Learning and Diverse Sampling for Lifelong Task-Oriented Dialogue Systems

被引:0
作者
Xu, Qiancheng [1 ]
Yang, Min [2 ]
Xu, Ruifeng [3 ]
机构
[1] Georgia Inst Technol, Atlanta, GA USA
[2] Chinese Acad Sci, Shenzhen Inst Adv Technol, Beijing, Peoples R China
[3] Harbin Inst Technol Shenzhen, Shenzhen, Peoples R China
来源
THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11 | 2023年
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In real-world scenarios, it is crucial to build a lifelong task-oriented dialogue system (TDS) that continually adapts to new knowledge without forgetting previously acquired experiences. Existing approaches mainly focus on mitigating the catastrophic forgetting in lifelong TDS. However, the transfer ability to generalize the accumulated old knowledge to new tasks is underexplored. In this paper, we propose a two-stage lifelong task-oriented dialogue generation method to mitigate catastrophic forgetting and encourage knowledge transfer simultaneously, inspired by the learning process. In the first stage, we learn task-specific masks which adaptively preserve the knowledge of each visited task so as to miti-gate catastrophic forgetting. In this stage, we are expected to learn the task-specific knowledge which is tailored for each task. In the second stage, we bring the knowledge from the encountered tasks together and understand thoroughly. To this end, we devise a balanced meta learning strategy for both forward and backward knowledge transfer in the lifelong learning process. In particular, we perform meta-update with a meta-test set sampled from the current training data for forward knowledge transfer. In addition, we em-ploy an uncertainty-based sampling strategy to select and store representative dialogue samples into episodic memory and perform meta-update with a meta-test set sampled from the memory for backward knowledge transfer. With extensive experiments on 29 tasks, we show that MetaLTDS out-performs the strong baselines in terms of both effectiveness and efficiency. For reproducibility, we submit our code at: https://github.com/travis-xu/MetaLTDS.
引用
收藏
页码:13843 / 13852
页数:10
相关论文
共 41 条
  • [21] Learning without Forgetting
    Li, Zhizhong
    Hoiem, Derek
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (12) : 2935 - 2947
  • [22] Lopez-Paz D, 2017, ADV NEUR IN, V30
  • [23] Madotto A, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P7452
  • [24] PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning
    Mallya, Arun
    Lazebnik, Svetlana
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7765 - 7773
  • [25] Continual Learning Dialogue Systems - Learning during Conversation
    Mazumder, Sahisnu
    Liu, Bing
    [J]. PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 3429 - 3432
  • [26] McCloskey M., 1989, PSYCHOL LEARN MOTIV, V24, P109
  • [27] The stability-plasticity dilemma: investigating the continuum from catastrophic forgetting to age-limited learning effects
    Mermillod, Martial
    Bugaiska, Aurelia
    Bonin, Patrick
    [J]. FRONTIERS IN PSYCHOLOGY, 2013, 4
  • [28] BLEU: a method for automatic evaluation of machine translation
    Papineni, K
    Roukos, S
    Ward, T
    Zhu, WJ
    [J]. 40TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, 2002, : 311 - 318
  • [29] SOLOIST: Building Task Bots at Scale with Transfer Learning and Machine Teaching
    Peng, Baolin
    Li, Chunyuan
    Li, Jinchao
    Shayandeh, Shahin
    Liden, Lars
    Gao, Jianfeng
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2021, 9 : 807 - 824
  • [30] Qu H, 2024, Arxiv, DOI arXiv:2109.11369