A Simple Multi-Armed Nearest-Neighbor Bandit for Interactive Recommendation

被引:23
|
作者
Sanz-Cruzado, Javier [1 ]
Castells, Pablo [1 ]
Lopez, Esther [1 ]
机构
[1] Univ Autonoma Madrid, Madrid, Spain
来源
RECSYS 2019: 13TH ACM CONFERENCE ON RECOMMENDER SYSTEMS | 2019年
关键词
Multi-armed bandits; Nearest-neighbors; Interactive recommendation; Thompson sampling;
D O I
10.1145/3298689.3347040
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The cyclic nature of the recommendation task is being increasingly taken into account in recommender systems research. In this line, framing interactive recommendation as a genuine reinforcement learning problem, multi-armed bandit approaches have been increasingly considered as a means to cope with the dual exploitation/exploration goal of recommendation. In this paper we develop a simple multi-armed bandit elaboration of neighbor-based collaborative filtering. The approach can be seen as a variant of the nearest-neighbors scheme, but endowed with a controlled stochastic exploration capability of the users' neighborhood, by a parameter-free application of Thompson sampling. Our approach is based on a formal development and a reasonably simple design, whereby it aims to be easy to reproduce and further elaborate upon. We report experiments using datasets from different domains showing that neighbor-based bandits indeed achieve recommendation accuracy enhancements in the mid to long run.
引用
收藏
页码:358 / 362
页数:5
相关论文
共 50 条
  • [31] Risk-averse Contextual Multi-armed Bandit Problem with Linear Payoffs
    Yifan Lin
    Yuhao Wang
    Enlu Zhou
    Journal of Systems Science and Systems Engineering, 2023, 32 : 267 - 288
  • [32] Risk-averse Contextual Multi-armed Bandit Problem with Linear Payoffs
    Lin, Yifan
    Wang, Yuhao
    Zhou, Enlu
    JOURNAL OF SYSTEMS SCIENCE AND SYSTEMS ENGINEERING, 2023, 32 (03) : 267 - 288
  • [33] Enhancing lane detection in autonomous vehicles with multi-armed bandit ensemble learning
    Pandian, J. Arun
    Thirunavukarasu, Ramkumar
    Mariappan, L. Thanga
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [34] Selecting multiple web adverts: A contextual multi-armed bandit with state uncertainty
    Edwards, James A.
    Leslie, David S.
    JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 2020, 71 (01) : 100 - 116
  • [35] Multi-Armed Bandit Beam Alignment and Tracking for Mobile Millimeter Wave Communications
    Booth, Matthew B.
    Suresh, Vinayak
    Michelusi, Nicolo
    Love, David J.
    IEEE COMMUNICATIONS LETTERS, 2019, 23 (07) : 1244 - 1248
  • [36] Influence Maximization Based Global Structural Properties: A Multi-Armed Bandit Approach
    Alshahrani, Mohammed
    Zhu Fuxi
    Sameh, Ahmed
    Mekouar, Soufiana
    Liu, Sichao
    IEEE ACCESS, 2019, 7 : 69707 - 69747
  • [37] Combinatorial Multi-Armed Bandit Based Unknown Worker Recruitment in Heterogeneous Crowdsensing
    Gao, Guoju
    Wu, Jie
    Xiao, Mingjun
    Chen, Guoliang
    IEEE INFOCOM 2020 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS, 2020, : 179 - 188
  • [38] Multi-Armed Bandit Algorithms for Crowdsourcing Systems with Online Estimation of Workers' Ability
    Rangi, Anshuka
    Franceschetti, Massimo
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 1345 - 1352
  • [39] Batch-Size Independent Regret Bounds for the Combinatorial Multi-Armed Bandit Problem
    Merlis, Nadav
    Mannor, Shie
    CONFERENCE ON LEARNING THEORY, VOL 99, 2019, 99
  • [40] FedAB: Truthful Federated Learning With Auction-Based Combinatorial Multi-Armed Bandit
    Wu, Chenrui
    Zhu, Yifei
    Zhang, Rongyu
    Chen, Yun
    Wang, Fangxin
    Cui, Shuguang
    IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (17) : 15159 - 15170