Balanced Meta Learning and Diverse Sampling for Lifelong Task-Oriented Dialogue Systems

被引:0
作者
Xu, Qiancheng [1 ]
Yang, Min [2 ]
Xu, Ruifeng [3 ]
机构
[1] Georgia Inst Technol, Atlanta, GA USA
[2] Chinese Acad Sci, Shenzhen Inst Adv Technol, Beijing, Peoples R China
[3] Harbin Inst Technol Shenzhen, Shenzhen, Peoples R China
来源
THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11 | 2023年
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In real-world scenarios, it is crucial to build a lifelong task-oriented dialogue system (TDS) that continually adapts to new knowledge without forgetting previously acquired experiences. Existing approaches mainly focus on mitigating the catastrophic forgetting in lifelong TDS. However, the transfer ability to generalize the accumulated old knowledge to new tasks is underexplored. In this paper, we propose a two-stage lifelong task-oriented dialogue generation method to mitigate catastrophic forgetting and encourage knowledge transfer simultaneously, inspired by the learning process. In the first stage, we learn task-specific masks which adaptively preserve the knowledge of each visited task so as to miti-gate catastrophic forgetting. In this stage, we are expected to learn the task-specific knowledge which is tailored for each task. In the second stage, we bring the knowledge from the encountered tasks together and understand thoroughly. To this end, we devise a balanced meta learning strategy for both forward and backward knowledge transfer in the lifelong learning process. In particular, we perform meta-update with a meta-test set sampled from the current training data for forward knowledge transfer. In addition, we em-ploy an uncertainty-based sampling strategy to select and store representative dialogue samples into episodic memory and perform meta-update with a meta-test set sampled from the memory for backward knowledge transfer. With extensive experiments on 29 tasks, we show that MetaLTDS out-performs the strong baselines in terms of both effectiveness and efficiency. For reproducibility, we submit our code at: https://github.com/travis-xu/MetaLTDS.
引用
收藏
页码:13843 / 13852
页数:10
相关论文
共 41 条
  • [1] Ahn H, 2019, ADV NEUR IN, V32
  • [2] Ba JL, 2016, arXiv
  • [3] Rainbow Memory: Continual Learning with a Memory of Diverse Samples
    Bang, Jihwan
    Kim, Heesu
    Yoo, YoungJoon
    Ha, Jung-Woo
    Choi, Jonghyun
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8214 - 8223
  • [4] Biesialska M., 2020, P INT C COMP LING, P6523, DOI 10.18653/v1/2020.coling-main.574
  • [5] Chaudhry A, 2019, Arxiv, DOI [arXiv:1812.00420, 10.48550/arXiv.1812.00420]
  • [6] Chaudhry Arslan, 2019, 7 INT C LEARNING REP
  • [7] d'Autume CD, 2019, Arxiv, DOI arXiv:1906.01076
  • [8] Gao TY, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P6894
  • [9] Geng BZ, 2021, ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, P517
  • [10] Goodfellow I., 2015, P INT C LEARN REPR