A Simple Multi-Armed Nearest-Neighbor Bandit for Interactive Recommendation

被引:23
|
作者
Sanz-Cruzado, Javier [1 ]
Castells, Pablo [1 ]
Lopez, Esther [1 ]
机构
[1] Univ Autonoma Madrid, Madrid, Spain
来源
RECSYS 2019: 13TH ACM CONFERENCE ON RECOMMENDER SYSTEMS | 2019年
关键词
Multi-armed bandits; Nearest-neighbors; Interactive recommendation; Thompson sampling;
D O I
10.1145/3298689.3347040
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The cyclic nature of the recommendation task is being increasingly taken into account in recommender systems research. In this line, framing interactive recommendation as a genuine reinforcement learning problem, multi-armed bandit approaches have been increasingly considered as a means to cope with the dual exploitation/exploration goal of recommendation. In this paper we develop a simple multi-armed bandit elaboration of neighbor-based collaborative filtering. The approach can be seen as a variant of the nearest-neighbors scheme, but endowed with a controlled stochastic exploration capability of the users' neighborhood, by a parameter-free application of Thompson sampling. Our approach is based on a formal development and a reasonably simple design, whereby it aims to be easy to reproduce and further elaborate upon. We report experiments using datasets from different domains showing that neighbor-based bandits indeed achieve recommendation accuracy enhancements in the mid to long run.
引用
收藏
页码:358 / 362
页数:5
相关论文
共 50 条
  • [21] Multi-Armed Bandit for Species Discovery: A Bayesian Nonparametric Approach
    Battiston, Marco
    Favaro, Stefano
    Teh, Yee Whye
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2018, 113 (521) : 455 - 466
  • [22] Mitigating starvation in dense WLANs: A multi-armed Bandit solution
    Bardou, Anthony
    Begin, Thomas
    Busson, Anthony
    AD HOC NETWORKS, 2023, 138
  • [23] Feedback Collection and Nearest-Neighbor Profiling for Recommendation Systems in Healthcare Scenarios
    Antonio, Joao
    Malheiro, Ricardo
    Jardim, Sandra
    TECHNOLOGIES, 2024, 12 (08)
  • [24] Multi-User Communication Networks: A Coordinated Multi-Armed Bandit Approach
    Avner, Orly
    Mannor, Shie
    IEEE-ACM TRANSACTIONS ON NETWORKING, 2019, 27 (06) : 2192 - 2207
  • [25] Randomized allocation with nonparametric estimation for a multi-armed bandit problem with covariates
    Yang, YH
    Zhu, D
    ANNALS OF STATISTICS, 2002, 30 (01) : 100 - 121
  • [26] Multi-Armed Bandit Algorithm Policy for LoRa Network Performance Enhancement
    Askhedkar, Anjali R.
    Chaudhari, Bharat S.
    JOURNAL OF SENSOR AND ACTUATOR NETWORKS, 2023, 12 (03)
  • [27] Thompson Sampling Based Mechanisms for Stochastic Multi-Armed Bandit Problems
    Ghalme, Ganesh
    Jain, Shweta
    Gujar, Sujit
    Narahari, Y.
    AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 87 - 95
  • [28] Multi-Armed Bandit-Based User Network Node Selection
    Gao, Qinyan
    Xie, Zhidong
    SENSORS, 2024, 24 (13)
  • [29] Unreliable Multi-Armed Bandits: A Novel Approach to Recommendation Systems
    Ravi, Aditya Narayan
    Poduval, Pranav
    Moharir, Sharayu
    2020 INTERNATIONAL CONFERENCE ON COMMUNICATION SYSTEMS & NETWORKS (COMSNETS), 2020,
  • [30] Risk-averse Contextual Multi-armed Bandit Problem with Linear Payoffs
    Lin, Yifan
    Wang, Yuhao
    Zhou, Enlu
    JOURNAL OF SYSTEMS SCIENCE AND SYSTEMS ENGINEERING, 2022,