State representation modeling for deep reinforcement learning based recommendation

被引:44
作者
Liu, Feng [1 ]
Tang, Ruiming [2 ]
Li, Xutao [1 ]
Zhang, Weinan [3 ]
Ye, Yunming [1 ]
Chen, Haokun [3 ]
Guo, Huifeng [2 ]
Zhang, Yuzhou [2 ]
He, Xiuqiang [2 ]
机构
[1] Harbin Inst Technol, Shenzhen Grad Sch, Shenzhen Key Lab Internet Informat Collaborat, Shenzhen 518055, Peoples R China
[2] Noahs Ark Lab, Huawei, Peoples R China
[3] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
基金
国家重点研发计划;
关键词
State representation modeling; Deep reinforcement learning; Recommendation;
D O I
10.1016/j.knosys.2020.106170
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning techniques have recently been introduced to interactive recommender systems to capture the dynamic patterns of user behavior during the interaction with recommender systems and perform planning to optimize long-term performance. Most existing research work focuses on designing policy and learning algorithms of the recommender agent but seldom cares about the state representation of the environment, which is indeed essential for the recommendation decision making. In this paper, we first formulate the interactive recommender system problem with a deep reinforcement learning recommendation framework. Within this framework, we then carefully design four state representation schemes for learning the recommendation policy. Inspired by recent advances in feature interaction modeling in user response prediction, we discover that explicitly modeling user-item interactions in state representation can largely help the recommendation policy perform effective reinforcement learning. Extensive experiments on four real-world datasets are conducted under both the offline and simulated online evaluation settings. The experimental results demonstrate the proposed state representation schemes lead to better performance over the state-of-the-art methods. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:12
相关论文
共 53 条
[21]   Field-aware Factorization Machines for CTR Prediction [J].
Juan, Yuchin ;
Zhuang, Yong ;
Chin, Wei-Sheng ;
Lin, Chih-Jen .
PROCEEDINGS OF THE 10TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'16), 2016, :43-50
[22]  
Koren Y., 2008, P 14 ACM SIGKDD INT, P426, DOI [DOI 10.1145/1401890.1401944, 10.1145/1401890.1401944]
[23]   MATRIX FACTORIZATION TECHNIQUES FOR RECOMMENDER SYSTEMS [J].
Koren, Yehuda ;
Bell, Robert ;
Volinsky, Chris .
COMPUTER, 2009, 42 (08) :30-37
[24]  
Li Lihong, 2010, P 19 INT C WORLD WID, P661, DOI DOI 10.1145/1772690.1772758
[25]   xDeepFM: Combining Explicit and Implicit Feature Interactions for Recommender Systems [J].
Lian, Jianxun ;
Zhou, Xiaohuan ;
Zhang, Fuzheng ;
Chen, Zhongxia ;
Xie, Xing ;
Sun, Guangzhong .
KDD'18: PROCEEDINGS OF THE 24TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2018, :1754-1763
[26]  
Lillicrap T.P., 2015, ARXIVABS150902971 CO
[27]   End-to-End Deep Reinforcement Learning based Recommendation with Supervised Embedding [J].
Liu, Feng ;
Guo, Huifeng ;
Li, Xutao ;
Tang, Ruiming ;
Ye, Yunming ;
He, Xiuqiang .
PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM '20), 2020, :384-392
[28]  
Maillard O.-A., 2011, Advances in Neural Information Processing Systems, V24, P2627
[29]   Multirelational Social Recommendations via Multigraph Ranking [J].
Mao, Mingsong ;
Lu, Jie ;
Zhang, Guangquan ;
Zhang, Jinlong .
IEEE TRANSACTIONS ON CYBERNETICS, 2017, 47 (12) :4049-4061
[30]  
McMahan HB, 2013, 19TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'13), P1222