A Dynamic Individual Recommendation Method Based on Reinforcement Learning

被引:0
作者
Han, Daojun [1 ]
Shen, Xiajiong [1 ]
Gan, Tian [2 ]
Cai, Ruiqing [3 ]
机构
[1] Henan Univ, Inst Data & Knowledge Engn, Kaifeng 475004, Henan, Peoples R China
[2] CITIC Bank Zhengzhou Branch, Zhengzhou 450000, Henan, Peoples R China
[3] Hikvis Digital Co Ltd, Hangzhou 310052, Zhejiang, Peoples R China
来源
PARALLEL ARCHITECTURE, ALGORITHM AND PROGRAMMING, PAAP 2017 | 2017年 / 729卷
基金
中国国家自然科学基金;
关键词
Recommendation method; Reinforcement learning; Dynamic;
D O I
10.1007/978-981-10-6442-5_17
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a widely used recommendation method, collaborative filtering can solve the problem of low level of resource utilization which caused by information overload. At present, in order to exhibiting and searching items, we need to use multipole attributes to describe items. Thus request to particularly distinguish every attribute and realize accurate recommendation. While the collaborative filtering method lose sight of the dynamic regulation of items attributes' importance degree, and it cannot interpose the discrimination of attributes. Aiming at this problem, this paper come up with a dynamic individual recommendation method based on reinforcement learning. This method can dig user's attribute tag preference from operant behavior. It can record user's attributes operate path and recall path. Then we build the award-punishment model of attribute tag, and realize the tag weight dynamic regulation. According to the principle that reinforcement learning system always get max award, we make a tag recommend strategy and give user recommendations in accordance with the preferences. The experimental result show that this method can distinguish the validity of user's click, and realize the tag weight dynamic regulation and give user recommendations in accordance with the preferences.
引用
收藏
页码:192 / 200
页数:9
相关论文
共 12 条
  • [1] Opportunities for multiagent systems and multiagent reinforcement learning in traffic control
    Bazzan, Ana L. C.
    [J]. AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2009, 18 (03) : 342 - 375
  • [2] Mining User Movement Behavior Patterns in a Mobile Service Environment
    Chen, Tzung-Shi
    Chou, Yen-Ssu
    Chen, Tzung-Cheng
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART A-SYSTEMS AND HUMANS, 2012, 42 (01): : 87 - 101
  • [3] Cong L., 2010, NEW TECHNOL LIB INF, V11, P7
  • [4] Gao Yang, 2004, Acta Automatica Sinica, V30, P86
  • [5] USING COLLABORATIVE FILTERING TO WEAVE AN INFORMATION TAPESTRY
    GOLDBERG, D
    NICHOLS, D
    OKI, BM
    TERRY, D
    [J]. COMMUNICATIONS OF THE ACM, 1992, 35 (12) : 61 - 70
  • [6] Reinforcement Learning strategies for A-Team solving the Resource-Constrained Project Scheduling Problem
    Jedrzejowicz, P.
    Ratajczak-Ropel, E.
    [J]. NEUROCOMPUTING, 2014, 146 : 301 - 307
  • [7] Reinforcement learning in robotics: A survey
    Kober, Jens
    Bagnell, J. Andrew
    Peters, Jan
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) : 1238 - 1274
  • [8] Li Hongtao, 2015, Journal of Sichuan University (Engineering Science Edition), V47, P131, DOI 10.15961/j.jsuese.2015.06.018
  • [9] [李全刚 Li Quangang], 2014, [计算机学报, Chinese Journal of Computers], V37, P1135
  • [10] Liu Z., 2015, COMPUT RES DEV, V03, P579