SEQ2SEQ++: A Multitasking-Based Seq2seq Model to Generate Meaningful and Relevant Answers

被引:6
|
作者
Palasundram, Kulothunkan [1 ]
Sharef, Nurfadhlina Mohd [1 ]
Kasmiran, Khairul Azhar [1 ]
Azman, Azreen [1 ]
机构
[1] Univ Putra Malaysia, Fac Comp Sci & Informat Technol, Intelligent Comp Res Grp, Seri Kembangan 43400, Selangor, Malaysia
来源
IEEE ACCESS | 2021年 / 9卷 / 09期
关键词
Task analysis; Chatbots; Computational modeling; Decoding; Training; Transformers; Benchmark testing; Sequence to sequence learning; natural answer generation; multitask learning; attention mechanism; ATTENTION; ENCODER;
D O I
10.1109/ACCESS.2021.3133495
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Question-answering chatbots have tremendous potential to complement humans in various fields. They are implemented using either rule-based or machine learning-based systems. Unlike the former, machine learning-based chatbots are more scalable. Sequence-to-sequence (Seq2Seq) learning is one of the most popular approaches in machine learning-based chatbots and has shown remarkable progress since its introduction in 2014. However, chatbots based on Seq2Seq learning show a weakness in that it tends to generate answers that can be generic and inconsistent with the questions, thereby becoming meaningless and, therefore, may lower the chatbot adoption rate. This weakness can be attributed to three issues: question encoder overfit, answer generation overfit, and language model influence. Several recent methods utilize multitask learning (MTL) to address this weakness. However, the existing MTL models show very little improvement over single-task learning, wherein they still generate generic and inconsistent answers. This paper presents a novel approach to MTL for the Seq2Seq learning model called SEQ2SEQ++, which comprises a multifunctional encoder, an answer decoder, an answer encoder, and a ternary classifier. Additionally, SEQ2SEQ++ utilizes a dynamic tasks loss weight mechanism for MTL loss calculation and a novel attention mechanism called the comprehensive attention mechanism. Experiments on NarrativeQA and SQuAD datasets were conducted to gauge the performance of the proposed model in comparison with two recently proposed models. The experimental results show that SEQ2SEQ++ yields noteworthy improvements over the two models on bilingual evaluation understudy, word error rate, and Distinct-2 metrics.
引用
收藏
页码:164949 / 164975
页数:27
相关论文
共 50 条
  • [41] PAMA-TTS: PROGRESSION-AWARE MONOTONIC ATTENTION FOR STABLE SEQ2SEQ TTS WITH ACCURATE PHONEME DURATION CONTROL
    He, Yunchao
    Luan, Jian
    Wang, Yujun
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7467 - 7471
  • [42] Short-time multi-energy load forecasting method based on CNN-Seq2Seq model with attention mechanism
    Zhang, Ge
    Bai, Xiaoqing
    Wang, Yuxuan
    MACHINE LEARNING WITH APPLICATIONS, 2021, 5
  • [43] Forcing-Seq2Seq Model: An Automatic Model of Title Generation for Natural Text Using Deep Learning
    Thuan Nguyen Thi Hiep
    Nhan To Thanh
    Tho Quan Thanh
    PROCEEDINGS OF THE FUTURE TECHNOLOGIES CONFERENCE (FTC) 2021, VOL 2, 2022, 359 : 388 - 402
  • [44] A general deep learning framework for history-dependent response prediction based on UA-Seq2Seq model
    Wang, Chen
    Xu, Li-yan
    Fan, Jian-sheng
    COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2020, 372
  • [45] AN ATTENTION-SEQ2SEQ MODEL BASED ON CRNN ENCODING FOR AUTOMATIC LABANOTATION GENERATION FROM MOTION CAPTURE DATA
    Li, Min
    Miao, Zhenjiang
    Zhang, Xiao-Ping
    Xu, Wanru
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4185 - 4189
  • [46] DA-CBGRU-Seq2Seq Based Soft Sensor for Mechanical Properties of Hot Rolling Process
    Liu, Xiaoyu
    Cong, Zhiyu
    Peng, Kaixiang
    Dong, Jie
    Li, Linlin
    IEEE SENSORS JOURNAL, 2023, 23 (13) : 14234 - 14244
  • [47] A novel algorithm system for wind power prediction based on RANSAC data screening and Seq2Seq-Attention-BiGRU model
    Zhou, Gaoyu
    Hu, Guofeng
    Zhang, Daxing
    Zhang, Yun
    ENERGY, 2023, 283
  • [48] Optimized EWT-Seq2Seq-LSTM with Attention Mechanism to Insulators Fault Prediction
    Klaar, Anne Carolina Rodrigues
    Stefenon, Stefano Frizzo
    Seman, Laio Oriel
    Mariani, Viviana Cocco
    Coelho, Leandro dos Santos
    SENSORS, 2023, 23 (06)
  • [49] MULTI-STEP WIND SPEED PREDICTION ALGORITHM OF MULTIPLE WIND TURBINES BASED ON SPATIAL-TEMPORAL ATTENTION-SEQ2SEQ MODEL
    Liu C.
    Zhao L.
    Wang Z.
    Xu J.
    Taiyangneng Xuebao/Acta Energiae Solaris Sinica, 2023, 44 (08): : 420 - 429
  • [50] A TMA-Seq2seq Network for Multi-Factor Time Series Sea Surface Temperature Prediction
    He, Qi
    Li, Wenlong
    Hao, Zengzhou
    Liu, Guohua
    Huang, Dongmei
    Song, Wei
    Xu, Huifang
    Alqahtani, Fayez
    Kim, Jeong-Uk
    CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 73 (01): : 51 - 67