Deep Reinforcement Learning Framework for Category-Based Item Recommendation

被引:24
作者
Fu, Mingsheng [1 ,2 ]
Agrawal, Anubha [3 ]
Irissappane, Athirai A. [3 ]
Zhang, Jie [4 ]
Huang, Liwei [1 ]
Qu, Hong [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 610054, Peoples R China
[2] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
[3] Univ Washington, Sch Engn & Technol, Tacoma, WA 98402 USA
[4] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
基金
中国博士后科学基金; 美国国家科学基金会;
关键词
Recommender systems; Reinforcement learning; Cybernetics; Computer science; Cats; Training; Research and development; Deep reinforcement learning (DRL); hierarchy; large action space; recommender system;
D O I
10.1109/TCYB.2021.3089941
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep reinforcement learning (DRL)-based recommender systems have recently come into the limelight due to their ability to optimize long-term user engagement. A significant challenge in DRL-based recommender systems is the large action space required to represent a variety of items. The large action space weakens the sampling efficiency and thereby, affects the recommendation accuracy. In this article, we propose a DRL-based method called deep hierarchical category-based recommender system (DHCRS) to handle the large action space problem. In DHCRS, categories of items are used to reconstruct the original flat action space into a two-level category-item hierarchy. DHCRS uses two deep Q-networks (DQNs): 1) a high-level DQN for selecting a category and 2) a low-level DQN to choose an item in this category for the recommendation. Hence, the action space of each DQN is significantly reduced. Furthermore, the categorization of items helps capture the users' preferences more effectively. We also propose a bidirectional category selection (BCS) technique, which explicitly considers the category-item relationships. The experiments show that DHCRS can significantly outperform state-of-the-art methods in terms of hit rate and normalized discounted cumulative gain for long-term recommendations.
引用
收藏
页码:12028 / 12041
页数:14
相关论文
共 37 条
  • [1] [Anonymous], 2015, Deep Reinforcement Learning in Large Discrete Action Spaces
  • [2] Bai X., 2019, ADV NEURAL INFORM PR, P10735
  • [3] Chen HK, 2019, AAAI CONF ARTIF INTE, P3312
  • [4] Top-K Off-Policy Correction for a REINFORCE Recommender System
    Chen, Minmin
    Beutel, Alex
    Covington, Paul
    Jain, Sagar
    Belletti, Francois
    Chi, Ed H.
    [J]. PROCEEDINGS OF THE TWELFTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM'19), 2019, : 456 - 464
  • [5] Chen XS, 2019, 36 INT C MACHINE LEA, V97
  • [6] A Novel Deep Learning-Based Collaborative Filtering Model for Recommendation System
    Fu, Mingsheng
    Qu, Hong
    Yi, Zhang
    Lu, Li
    Liu, Yongsheng
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2019, 49 (03) : 1084 - 1096
  • [7] Fujimoto S, 2019, PR MACH LEARN RES, V97
  • [8] DRCGR: Deep reinforcement learning framework incorporating CNN and GAN-based for interactive recommendation
    Gao, Rong
    Xia, Haifeng
    Li, Jing
    Liu, Donghua
    Chen, Shuai
    Chun, Gang
    [J]. 2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019), 2019, : 1048 - 1053
  • [9] He J, 2017, PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P1837
  • [10] Neural Collaborative Filtering
    He, Xiangnan
    Liao, Lizi
    Zhang, Hanwang
    Nie, Liqiang
    Hu, Xia
    Chua, Tat-Seng
    [J]. PROCEEDINGS OF THE 26TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB (WWW'17), 2017, : 173 - 182