Learning From Atypical Behavior: Temporary Interest Aware Recommendation Based on Reinforcement Learning

被引:5
作者
Du, Ziwen [1 ]
Yang, Ning [1 ]
Yu, Zhonghua [1 ]
Yu, Philip S. [2 ]
机构
[1] Sichuan Univ, Sch Comp Sci, Chengdu 610017, Peoples R China
[2] Univ Illinois, Dept Comp Sci, Chicago 60607, IL USA
基金
中国国家自然科学基金;
关键词
Robust recommendation; temporary interest; deep reinforcement learning;
D O I
10.1109/TKDE.2022.3144292
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traditional robust recommendation methods view atypical user-item interactions as noise and aim to reduce their impact with some kind of noise filtering technique, which often suffers from two challenges. First, in real world, atypical interactions may signal users' temporary interest different from their general preference. Therefore, simply filtering out the atypical interactions as noise may be inappropriate and degrade the personalization of recommendations. Second, it is hard to acquire the temporary interest since there are no explicit supervision signals to indicate whether an interaction is atypical or not. To address this challenges, we propose a novel model called Temporary Interest Aware Recommendation (TIARec), which can distinguish atypical interactions from normal ones without supervision and capture the temporary interest as well as the general preference of users. Particularly, we propose a reinforcement learning framework containing a recommender agent and an auxiliary classifier agent, which are jointly trained with the objective of maximizing the cumulative return of the recommendations made by the recommender agent. During the joint training process, the classifier agent can judge whether the interaction with an item recommended by the recommender agent is atypical, and the knowledge about learning temporary interest from atypical interactions can be transferred to the recommender agent, which makes the recommender agent able to alone make recommendations that balance the general preference and temporary interest of users. At last, the experiments conducted on real world datasets verify the effectiveness of TIARec.
引用
收藏
页码:9824 / 9835
页数:12
相关论文
共 36 条
[1]  
[Anonymous], 2008, P 25 INT C MACH LEAR, DOI DOI 10.1145/1390156.1390294
[2]  
[Anonymous], 2012, CoRR
[3]  
Chen HK, 2019, AAAI CONF ARTIF INTE, P3312
[4]   Top-K Off-Policy Correction for a REINFORCE Recommender System [J].
Chen, Minmin ;
Beutel, Alex ;
Covington, Paul ;
Jain, Sagar ;
Belletti, Francois ;
Chi, Ed H. .
PROCEEDINGS OF THE TWELFTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM'19), 2019, :456-464
[5]   How Dataset Characteristics Affect the Robustness of Collaborative Recommendation Models [J].
Deldjoo, Yashar ;
Di Noia, Tommaso ;
Di Sciascio, Eugenio ;
Antonio, Felice .
PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, :951-960
[6]  
Dulac-Arnold G, 2016, Arxiv, DOI arXiv:1512.07679
[7]   Learning to Collaborate: Multi-Scenario Ranking via Multi-Agent Reinforcement Learning [J].
Feng, Jun ;
Li, Heng ;
Huang, Minlie ;
Liu, Shichen ;
Ou, Wenwu ;
Wang, Zhirong ;
Zhu, Xiaoyan .
WEB CONFERENCE 2018: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW2018), 2018, :1939-1948
[8]   Mention Recommendation in Twitter with Cooperative Multi-Agent Reinforcement Learning [J].
Gui, Tao ;
Liu, Peng ;
Zhang, Qi ;
Zhu, Liang ;
Peng, Minlong ;
Zhou, Yunhua ;
Huang, Xuanjing .
PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, :535-544
[9]   Adversarial Personalized Ranking for Recommendation [J].
He, Xiangnan ;
He, Zhankui ;
Du, Xiaoyu ;
Chua, Tat-Seng .
ACM/SIGIR PROCEEDINGS 2018, 2018, :355-364
[10]   Learning to Collaborate in Multi-Module Recommendation via Multi-Agent Reinforcement Learning without Communication [J].
He, Xu ;
An, Bo ;
Li, Yanghua ;
Chen, Haikai ;
Wang, Rundong ;
Wang, Xinrun ;
Yu, Runsheng ;
Li, Xin ;
Wang, Zhirong .
RECSYS 2020: 14TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, 2020, :210-219