Learning Relevant Questions for Conversational Product Search using Deep Reinforcement Learning

被引:2
|
作者
Montazeralghaem, Ali [1 ]
Allan, James [1 ]
机构
[1] Univ Massachusetts, Ctr Intelligent Informat Retrieval, Coll Informat & Comp Sci, Amherst, MA 01003 USA
来源
WSDM'22: PROCEEDINGS OF THE FIFTEENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING | 2022年
关键词
Conversational Product Search; Reinforcement Learning; Relevant Questions; Intelligent Assistants; GAME; GO;
D O I
10.1145/3488560.3498526
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose RelQuest, a conversational product search model based on reinforcement learning to generate questions from product descriptions in each round of the conversation, directly maximizing any desired metrics (i.e., the ultimate goal of the conversation), objectives, or even an arbitrary user satisfaction signal. By enabling systems to ask questions about user needs, conversational product search has gained increasing attention in recent years. Asking the right questions through conversations helps the system collect valuable feedback to create better user experiences and ultimately increase sales. In contrast, existing conversational product search methods are based on an assumption that there is a set of effectively pre-defined candidate questions for each product to be asked. Moreover, they make strong assumptions to estimate the value of questions in each round of the conversation. Estimating the true value of questions in each round of the conversation is not trivial since it is unknown. Experiments on real-world user purchasing data show the effectiveness of RelQuest at generating questions that maximize standard evaluation measures such as NDCG.
引用
收藏
页码:746 / 754
页数:9
相关论文
共 50 条
  • [41] Reinforcement Learning to Create Value and Policy Functions Using Minimax Tree Search in Hex
    Takada, Kei
    Iizuka, Hiroyuki
    Yamamoto, Masahito
    IEEE TRANSACTIONS ON GAMES, 2020, 12 (01) : 63 - 73
  • [42] Deep reinforcement learning for time-critical wilderness search and rescue using drones
    Ewers, Jan-Hendrik
    Anderson, David
    Thomson, Douglas
    FRONTIERS IN ROBOTICS AND AI, 2025, 11
  • [43] Transfer Learning in Deep Reinforcement Learning
    Islam, Tariqul
    Abid, Dm. Mehedi Hasan
    Rahman, Tanvir
    Zaman, Zahura
    Mia, Kausar
    Hossain, Ramim
    PROCEEDINGS OF SEVENTH INTERNATIONAL CONGRESS ON INFORMATION AND COMMUNICATION TECHNOLOGY, ICICT 2022, VOL 1, 2023, 447 : 145 - 153
  • [44] A deep reinforcement learning approach for chemical production scheduling
    Hubbs, Christian D.
    Li, Can
    Sahinidis, Nikolaos, V
    Grossmann, Ignacio E.
    Wassick, John M.
    COMPUTERS & CHEMICAL ENGINEERING, 2020, 141
  • [45] Optimization of global production scheduling with deep reinforcement learning
    Waschneck, Bernd
    Reichstaller, Andre
    Belzner, Lenz
    Altenmueller, Thomas
    Bauernhansl, Thomas
    Knapp, Alexander
    Kyek, Andreas
    51ST CIRP CONFERENCE ON MANUFACTURING SYSTEMS, 2018, 72 : 1264 - 1269
  • [46] Reinforcement learning for control: Performance, stability, and deep approximators
    Busoniu, Lucian
    de Bruin, Tim
    Tolic, Domagoj
    Kober, Jens
    Palunko, Ivana
    ANNUAL REVIEWS IN CONTROL, 2018, 46 : 8 - 28
  • [47] RLCFR: Minimize counterfactual regret by deep reinforcement learning
    Li, Huale
    Wang, Xuan
    Jia, Fengwei
    Wu, Yulin
    Zhang, Jiajia
    Qi, Shuhan
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 187
  • [48] A Survey on Population-Based Deep Reinforcement Learning
    Long, Weifan
    Hou, Taixian
    Wei, Xiaoyi
    Yan, Shichao
    Zhai, Peng
    Zhang, Lihua
    MATHEMATICS, 2023, 11 (10)
  • [49] Road Planning for Slums via Deep Reinforcement Learning
    Zheng, Yu
    Su, Hongyuan
    Ding, Jingtao
    Jin, Depeng
    Li, Yong
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 5695 - 5706
  • [50] Hierarchical Deep Reinforcement Learning for Continuous Action Control
    Yang, Zhaoyang
    Merrick, Kathryn
    Jin, Lianwen
    Abbass, Hussein A.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (11) : 5174 - 5184