Multi-modal multi-hop interaction network for dialogue response generation

被引:4
|
作者
Zhou, Jie [1 ,2 ]
Tian, Junfeng [3 ]
Wang, Rui [4 ]
Wu, Yuanbin [2 ]
Yan, Ming [3 ]
He, Liang [1 ,2 ]
Huang, Xuanjing [5 ]
机构
[1] East China Normal Univ, Shanghai Key Lab Multidimens Informat Proc, Shanghai, Peoples R China
[2] East China Normal Univ, Dept Comp Sci & Technol, Shanghai, Peoples R China
[3] Alibaba Grp, Hangzhou, Peoples R China
[4] Vipshop China Co Ltd, Guangzhou, Peoples R China
[5] Fudan Univ, Sch Comp Sci, Shanghai, Peoples R China
关键词
Dialogue response generation; Multimodal; Interaction;
D O I
10.1016/j.eswa.2023.120267
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most task-oriented dialogue systems generate informative and appropriate responses by leveraging structured knowledge bases which, in practise, are not always available. For instance, in the e-commerce scenario, commercial items often miss key attribute values, while containing abundant unstructured multi-modal information, e.g., text description and images. Previous studies have not fully explored such information for dialogue response generation. In this paper, we propose a Multi-modal multi-hop Interaction Network for Dialogue (MIND) to facilitate 1) the interaction between a query and multi-modal information through the query-aware multi-modal encoder and 2) the interaction between modalities through the multi-hop decoder. We conduct extensive experiments to demonstrate the effectiveness of MIND over strong baselines, which achieves state-of-the-art performance for automatic and human evaluation. We also release two real-world large-scale datasets containing both dialogue history and items' multi-modal information to facilitate future research.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Increasing the Network Capacity for Multi-modal Multi-hop WSNs through Unsupervised Data Rate Adjustment
    Jones, Matthew
    Bein, Doina
    Madan, Bharat B.
    Phoha, Shashi
    INTELLIGENT DISTRIBUTED COMPUTING V, 2011, 382 : 183 - +
  • [2] Enhancing Multi-modal Multi-hop Question Answering via Structured Knowledge and Unified Retrieval-Generation
    Yang, Qian
    Chen, Qian
    Wang, Wen
    Hu, Baotian
    Zhang, Min
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5223 - 5234
  • [3] Multi-Modal Alignment of Visual Question Answering Based on Multi-Hop Attention Mechanism
    Xia, Qihao
    Yu, Chao
    Hou, Yinong
    Peng, Pingping
    Zheng, Zhengqi
    Chen, Wen
    ELECTRONICS, 2022, 11 (11)
  • [4] A Multi-Hop Reasoning Knowledge Selection Module for Dialogue Generation
    Ma, Zhiqiang
    Liu, Jia
    Xu, Biqi
    Lv, Kai
    Guo, Siyuan
    ELECTRONICS, 2024, 13 (16)
  • [5] Multi-hop neighbor fusion enhanced hierarchical transformer for multi-modal knowledge graph completion
    Wang, Yunpeng
    Ning, Bo
    Wang, Xin
    Li, Guanyu
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2024, 27 (05):
  • [6] Multi-hop Question Generation with Graph Convolutional Network
    Su, Dan
    Xu, Yan
    Dai, Wenliang
    Ji, Ziwei
    Yu, Tiezheng
    Fung, Pascale
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 4636 - 4647
  • [7] Architecture of multi-modal dialogue system
    Fuchs, M
    Hejda, P
    Slavík, P
    TEXT, SPEECH AND DIALOGUE, PROCEEDINGS, 2000, 1902 : 433 - 438
  • [8] Discovering Multimodal Hierarchical Structures with Graph Neural Networks for Multi-modal and Multi-hop Question Answering
    Zhang, Qing
    Lv, Haocheng
    Liu, Jie
    Chen, Zhiyun
    Duan, Jianyong
    Xv, Mingying
    Wang, Hao
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT I, 2024, 14425 : 383 - 394
  • [9] Multi-modal human robot interaction for map generation
    Saito, H
    Ishimura, K
    Hattori, M
    Takamori, T
    SICE 2002: PROCEEDINGS OF THE 41ST SICE ANNUAL CONFERENCE, VOLS 1-5, 2002, : 2721 - 2724
  • [10] Multi-level Interaction Network for Multi-Modal Rumor Detection
    Zou, Ting
    Qian, Zhong
    Li, Peifeng
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,