ReBoost: a retrieval-boosted sequence-to-sequence model for neural response generation

被引:6
|
作者
Zhu, Yutao [1 ,2 ]
Dou, Zhicheng [1 ,2 ]
Nie, Jian-Yun [3 ]
Wen, Ji-Rong [1 ,2 ]
机构
[1] Renmin Univ China, Sch Informat, Beijing, Peoples R China
[2] Beijing Key Lab Big Data Management & Anal Method, Beijing, Peoples R China
[3] Univ Montreal, DIRO, CP 6128,Succ Ctr Ville, Montreal, PQ, Canada
来源
INFORMATION RETRIEVAL JOURNAL | 2020年 / 23卷 / 01期
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Retrieved results; Seq2seq model; Response generation;
D O I
10.1007/s10791-019-09364-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Human-computer conversation is an active research topic in natural language processing. One of the representative methods to build conversation systems uses the sequence-to-sequence (Seq2seq) model through neural networks. However, with limited input information, the Seq2seq model tends to generate meaningless and trivial responses. It can be greatly enhanced if more supplementary information is provided in the generation process. In this work, we propose to utilize retrieved responses to boost the Seq2seq model for generating more informative replies. Our method, called ReBoost, incorporates retrieved results in the Seq2seq model by a hierarchical structure. The input message and retrieved results can influence the generation process jointly. Experiments on two benchmark datasets demonstrate that our model is able to generate more informative responses in both automatic and human evaluations and outperforms the state-of-the-art response generation models.
引用
收藏
页码:27 / 48
页数:22
相关论文
共 50 条
  • [21] Persian Keyphrase Generation Using Sequence-to-sequence Models
    Doostmohammadi, Ehsan
    Bokaei, Mohammad Hadi
    Sameti, Hossein
    2019 27TH IRANIAN CONFERENCE ON ELECTRICAL ENGINEERING (ICEE 2019), 2019, : 2010 - 2015
  • [22] A Sequence-to-Sequence Model for Semantic Role Labeling
    Daza, Angel
    Frank, Anette
    REPRESENTATION LEARNING FOR NLP, 2018, : 207 - 216
  • [23] Document Ranking with a Pretrained Sequence-to-Sequence Model
    Nogueira, Rodrigo
    Jiang, Zhiying
    Pradeep, Ronak
    Lin, Jimmy
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 708 - 718
  • [24] INVESTIGATION OF AN INPUT SEQUENCE ON THAI NEURAL SEQUENCE-TO-SEQUENCE SPEECH SYNTHESIS
    Janyoi, Pongsathon
    Thangthai, Ausdang
    2021 24TH CONFERENCE OF THE ORIENTAL COCOSDA INTERNATIONAL COMMITTEE FOR THE CO-ORDINATION AND STANDARDISATION OF SPEECH DATABASES AND ASSESSMENT TECHNIQUES (O-COCOSDA), 2021, : 218 - 223
  • [25] Jointly Masked Sequence-to-Sequence Model for Non-Autoregressive Neural Machine Translation
    Guo, Junliang
    Xu, Linli
    Chen, Enhong
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 376 - 385
  • [26] A Sequence-to-Sequence Transformer Model for Satellite Retrieval of Aerosol Optical and Microphysical Parameters from Space
    Zhang, Luo
    Gu, Haoran
    Li, Zhengqiang
    Liu, Zhenhai
    Zhang, Ying
    Xie, Yisong
    Zhang, Zihan
    Ji, Zhe
    Li, Zhiyu
    Yan, Chaoyu
    REMOTE SENSING, 2024, 16 (24)
  • [27] Automatic Generation of Artificial Space Weather Forecast Product Based on Sequence-to-sequence Model
    罗冠霆
    ZOU Yenan
    CAI Yanxia
    空间科学学报, 2024, 44 (01) : 80 - 94
  • [28] De-duping URLs with Sequence-to-Sequence Neural Networks
    Xu, Keyang
    Liu, Zhengzhong
    Callan, Jamie
    SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, : 1157 - 1160
  • [29] Retrosynthetic Reaction Prediction Using Neural Sequence-to-Sequence Models
    Liu, Bowen
    Ramsundar, Bharath
    Kawthekar, Prasad
    Shi, Jade
    Gomes, Joseph
    Quang Luu Nguyen
    Ho, Stephen
    Sloane, Jack
    Wender, Paul
    Pande, Vijay
    ACS CENTRAL SCIENCE, 2017, 3 (10) : 1103 - 1113
  • [30] SEQUENCE-TO-SEQUENCE LABANOTATION GENERATION BASED ON MOTION CAPTURE DATA
    Li, Min
    Miao, Zhenjiang
    Ma, Cong
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 4517 - 4521