Knowledge-Grounded Dialogue Generation with Pre-trained Language Models

被引:0
作者
Zhao, Xueliang [1 ,2 ]
Wu, Wei [3 ]
Xu, Can [4 ]
Tao, Chongyang [4 ]
Zhao, Dongyan [1 ,2 ]
Yan, Rui [1 ,2 ,5 ]
机构
[1] Peking Univ, Wangxuan Inst Comp Technol, Beijing, Peoples R China
[2] Peking Univ, Ctr Data Sci, AAIS, Beijing, Peoples R China
[3] Meituan, Beijing, Peoples R China
[4] Microsoft Corp, Beijing, Peoples R China
[5] Beijing Acad Artificial Intelligence BAAI, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP) | 2020年
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study knowledge-grounded dialogue generation with pre-trained language models. To leverage the redundant external knowledge under capacity constraint, we propose equipping response generation defined by a pre-trained language model with a knowledge selection module, and an unsupervised approach to jointly optimizing knowledge selection and response generation with unlabeled dialogues. Empirical results on two benchmarks indicate that our model can significantly outperform state-of-the-art methods in both automatic evaluation and human judgment.
引用
收藏
页码:3377 / 3390
页数:14
相关论文
共 50 条
  • [21] Adaptive Posterior Knowledge Selection for Improving Knowledge-Grounded Dialogue Generation
    Wang, Weichao
    Gao, Wei
    Feng, Shi
    Chen, Ling
    Wang, Daling
    [J]. International Conference on Information and Knowledge Management, Proceedings, 2021, : 1989 - 1998
  • [22] Probing Simile Knowledge from Pre-trained Language Models
    Chen, Weijie
    Chang, Yongzhu
    Zhang, Rongsheng
    Pu, Jiashu
    Chen, Guandan
    Zhang, Le
    Xi, Yadong
    Chen, Yijiang
    Su, Chang
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5875 - 5887
  • [23] Efficient Latent Variable Modeling for Knowledge-Grounded Dialogue Generation
    Han, Gunsoo
    Jo, Daejin
    Nam, Daniel Wontae
    Yoon, Eunseop
    Kwon, Taehwan
    Rho, Seungeun
    On, Kyoung-Woon
    Yoo, Chang D.
    Kim, Sungwoong
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 2683 - 2702
  • [24] ProSide: Knowledge Projector and Sideway for Pre-trained Language Models
    He, Chaofan
    Lu, Gewei
    Shen, Liping
    [J]. NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT II, NLPCC 2024, 2025, 15360 : 56 - 68
  • [25] Continual knowledge infusion into pre-trained biomedical language models
    Jha, Kishlay
    Zhang, Aidong
    [J]. BIOINFORMATICS, 2022, 38 (02) : 494 - 502
  • [26] Exploring Pre-trained Language Models for Event Extraction and Generation
    Yang, Sen
    Feng, Dawei
    Qiao, Linbo
    Kan, Zhigang
    Li, Dongsheng
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5284 - 5294
  • [27] STYLEDGPT: Stylized Response Generation with Pre-trained Language Models
    Yang, Ze
    Wu, Wei
    Xu, Can
    Liang, Xinnian
    Bai, Jiaqi
    Wang, Liran
    Wang, Wei
    Li, Zhoujun
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1548 - 1559
  • [28] Scalable Educational Question Generation with Pre-trained Language Models
    Bulathwela, Sahan
    Muse, Hamze
    Yilmaz, Emine
    [J]. ARTIFICIAL INTELLIGENCE IN EDUCATION, AIED 2023, 2023, 13916 : 327 - 339
  • [29] TopicKS: Topic-driven Knowledge Selection for Knowledge-grounded Dialogue Generation
    Wang, Shiquan
    Si, Yuke
    Wei, Xiao
    Wang, Longbiao
    Zhuang, Zhiqiang
    Zhang, Xiaowang
    Dang, Jianwu
    [J]. INTERSPEECH 2022, 2022, : 1121 - 1125
  • [30] Pre-Trained Language Models and Their Applications
    Wang, Haifeng
    Li, Jiwei
    Wu, Hua
    Hovy, Eduard
    Sun, Yu
    [J]. ENGINEERING, 2023, 25 : 51 - 65