ReBoost: a retrieval-boosted sequence-to-sequence model for neural response generation

被引:6
|
作者
Zhu, Yutao [1 ,2 ]
Dou, Zhicheng [1 ,2 ]
Nie, Jian-Yun [3 ]
Wen, Ji-Rong [1 ,2 ]
机构
[1] Renmin Univ China, Sch Informat, Beijing, Peoples R China
[2] Beijing Key Lab Big Data Management & Anal Method, Beijing, Peoples R China
[3] Univ Montreal, DIRO, CP 6128,Succ Ctr Ville, Montreal, PQ, Canada
来源
INFORMATION RETRIEVAL JOURNAL | 2020年 / 23卷 / 01期
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Retrieved results; Seq2seq model; Response generation;
D O I
10.1007/s10791-019-09364-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Human-computer conversation is an active research topic in natural language processing. One of the representative methods to build conversation systems uses the sequence-to-sequence (Seq2seq) model through neural networks. However, with limited input information, the Seq2seq model tends to generate meaningless and trivial responses. It can be greatly enhanced if more supplementary information is provided in the generation process. In this work, we propose to utilize retrieved responses to boost the Seq2seq model for generating more informative replies. Our method, called ReBoost, incorporates retrieved results in the Seq2seq model by a hierarchical structure. The input message and retrieved results can influence the generation process jointly. Experiments on two benchmark datasets demonstrate that our model is able to generate more informative responses in both automatic and human evaluations and outperforms the state-of-the-art response generation models.
引用
收藏
页码:27 / 48
页数:22
相关论文
共 50 条
  • [31] AN ANALYSIS OF INCORPORATING AN EXTERNAL LANGUAGE MODEL INTO A SEQUENCE-TO-SEQUENCE MODEL
    Kannan, Anjuli
    Wu, Yonghui
    Nguyen, Patrick
    Sainath, Tara N.
    Chen, Zhifeng
    Prabhavalkar, Rohit
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 5824 - 5828
  • [32] A Dataset for Low-Resource Stylized Sequence-to-Sequence Generation
    Wu, Yu
    Wang, Yunli
    Liu, Shujie
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 9290 - 9297
  • [33] Understanding and Improving Sequence-to-Sequence Pretraining for Neural Machine Translation
    Wang, Wenxuan
    Jiao, Wenxiang
    Hao, Yongchang
    Wang, Xing
    Shi, Shuming
    Tu, Zhaopeng
    Lyu, Michael R.
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2591 - 2600
  • [34] Explainable sequence-to-sequence GRU neural network for pollution forecasting
    Sara Mirzavand Borujeni
    Leila Arras
    Vignesh Srinivasan
    Wojciech Samek
    Scientific Reports, 13
  • [35] Explainable sequence-to-sequence GRU neural network for pollution forecasting
    Borujeni, Sara Mirzavand
    Arras, Leila
    Srinivasan, Vignesh
    Samek, Wojciech
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [36] FPGA implementation of sequence-to-sequence predicting spiking neural networks
    Ye, ChangMin
    Kornijcuk, Vladimir
    Kim, Jeeson
    Jeong, Doo Seok
    2020 17TH INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC 2020), 2020, : 322 - 323
  • [37] A Sequence-to-Sequence Pronunciation Model for Bangla Speech Synthesis
    Ahmad, Arif
    Hussain, Mohammed Raihan
    Selim, Mohammad Reza
    Iqbal, Muhammed Zafar
    Rahman, Mohammad Shahidur
    2018 INTERNATIONAL CONFERENCE ON BANGLA SPEECH AND LANGUAGE PROCESSING (ICBSLP), 2018,
  • [38] Sequence-to-Sequence Model with Attention for Time Series Classification
    Tang, Yujin
    Xu, Jianfeng
    Matsumoto, Kazunori
    Ono, Chihiro
    2016 IEEE 16TH INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2016, : 503 - 510
  • [39] A Hierarchical Sequence-to-Sequence Model for Korean POS Tagging
    Jin, Guozhe
    Yu, Zhezhou
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2021, 20 (02)
  • [40] BARThez: a Skilled Pretrained French Sequence-to-Sequence Model
    Eddine, Moussa Kamal
    Tixier, Antoine J-P
    Vazirgiannis, Michalis
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 9369 - 9390