GLM-Dialog: Noise-tolerant Pre-training for Knowledge-grounded Dialogue Generation

被引:3
作者
Zhang, Jing [1 ]
Zhang, Xiaokang [1 ]
Zhang-Li, Daniel [2 ]
Yu, Jifan [2 ]
Yao, Zijun [2 ]
Ma, Zeyao [3 ]
Xu, Yiqi [1 ]
Wang, Haohua [2 ]
Zhang, Xiaohan [4 ]
Lin, Nianyi [2 ]
Lu, Sunrui [2 ]
Li, Juanzi [2 ]
Tang, Jie [2 ]
机构
[1] Renmin Univ China, Beijing, Peoples R China
[2] Tsinghua Univ, Beijing, Peoples R China
[3] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
[4] ZHIPU AI, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023 | 2023年
关键词
Dialogue System; Dialogue Evaluation; Large Language Model;
D O I
10.1145/3580305.3599832
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present GLM-Dialog, a large-scale language model (LLM) with 10B parameters capable of knowledge-grounded conversation in Chinese using a search engine to access the Internet knowledge. GLM-Dialog offers a series of applicable techniques for exploiting various external knowledge including both helpful and noisy knowledge, enabling the creation of robust knowledge-grounded dialogue LLMs with limited proper datasets. To evaluate the GLM-Dialog more fairly, we also propose a novel evaluation method to allow humans to converse with multiple deployed bots simultaneously and compare their performance implicitly instead of explicitly rating using multidimensional metrics. Comprehensive evaluations from automatic to human perspective demonstrate the advantages of GLM-Dialog comparing with existing open source Chinese dialogue models. We release both the model checkpoint and source code, and also deploy it as a WeChat application to interact with users(1). We offer our evaluation platform online(2) in an effort to prompt the development of open source models and reliable dialogue evaluation systems. All the source code is available on Github(3).
引用
收藏
页码:5564 / 5575
页数:12
相关论文
共 40 条
  • [1] Conversational agents for fostering curiosity-driven learning in children
    Abdelghani, Rania
    Oudeyer, Pierre-Yves
    Law, Edith
    de Vulpillieres, Catherine
    Sauzeon, Helene
    [J]. INTERNATIONAL JOURNAL OF HUMAN-COMPUTER STUDIES, 2022, 167
  • [2] [Anonymous], NAT COMMUN
  • [3] [Anonymous], **DATA OBJECT**, DOI DOI 10.5281/ZENODO.3402023
  • [4] [Anonymous], 2017, ARXIV171105073
  • [5] Bao Siqi, 2021, ARXIV210909519
  • [6] Bao Siqi, 2022, ARXIV221100910
  • [7] Cui Y., 2020, ABS200413922 CORR
  • [8] Dinan Emily, 2018, INT C LEARN REPR
  • [9] Dong L, 2019, ADV NEUR IN, V32
  • [10] Du ZX, 2022, PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), P320