A Simple Multi-Armed Nearest-Neighbor Bandit for Interactive Recommendation

被引:23
|
作者
Sanz-Cruzado, Javier [1 ]
Castells, Pablo [1 ]
Lopez, Esther [1 ]
机构
[1] Univ Autonoma Madrid, Madrid, Spain
来源
RECSYS 2019: 13TH ACM CONFERENCE ON RECOMMENDER SYSTEMS | 2019年
关键词
Multi-armed bandits; Nearest-neighbors; Interactive recommendation; Thompson sampling;
D O I
10.1145/3298689.3347040
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The cyclic nature of the recommendation task is being increasingly taken into account in recommender systems research. In this line, framing interactive recommendation as a genuine reinforcement learning problem, multi-armed bandit approaches have been increasingly considered as a means to cope with the dual exploitation/exploration goal of recommendation. In this paper we develop a simple multi-armed bandit elaboration of neighbor-based collaborative filtering. The approach can be seen as a variant of the nearest-neighbors scheme, but endowed with a controlled stochastic exploration capability of the users' neighborhood, by a parameter-free application of Thompson sampling. Our approach is based on a formal development and a reasonably simple design, whereby it aims to be easy to reproduce and further elaborate upon. We report experiments using datasets from different domains showing that neighbor-based bandits indeed achieve recommendation accuracy enhancements in the mid to long run.
引用
收藏
页码:358 / 362
页数:5
相关论文
共 50 条
  • [1] Multi-Armed Recommender System Bandit Ensembles
    Canamares, Rocio
    Redondo, Marcos
    Castells, Pablo
    RECSYS 2019: 13TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, 2019, : 432 - 436
  • [2] Scaling Multi-Armed Bandit Algorithms
    Fouche, Edouard
    Komiyama, Junpei
    Boehm, Klemens
    KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, : 1449 - 1459
  • [3] A Multi-Armed Bandit Model Selection for Cold-Start User Recommendation
    Felicio, Cricia Z.
    Paixao, Klerisson V. R.
    Barcelos, Celia A. Z.
    Preux, Philippe
    PROCEEDINGS OF THE 25TH CONFERENCE ON USER MODELING, ADAPTATION AND PERSONALIZATION (UMAP'17), 2017, : 32 - 40
  • [4] Characterizing Truthful Multi-Armed Bandit Mechanisms
    Babaioff, Moshe
    Sharma, Yogeshwer
    Slivkins, Aleksandrs
    10TH ACM CONFERENCE ON ELECTRONIC COMMERCE - EC 2009, 2009, : 79 - 88
  • [5] CHARACTERIZING TRUTHFUL MULTI-ARMED BANDIT MECHANISMS
    Babaioff, Moshe
    Sharma, Yogeshwer
    Slivkins, Aleksandrs
    SIAM JOURNAL ON COMPUTING, 2014, 43 (01) : 194 - 230
  • [6] A Bayesian Multi-armed Bandit Approach for Identifying Human Vulnerabilities
    Miehling, Erik
    Xiao, Baicen
    Poovendran, Radha
    Basar, Tamer
    DECISION AND GAME THEORY FOR SECURITY, GAMESEC 2018, 2018, 11199 : 521 - 539
  • [7] Achieving Fairness in the Stochastic Multi-Armed Bandit Problem
    Patil, Vishakha
    Ghalme, Ganesh
    Nair, Vineet
    Narahari, Y.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [8] Repeated Dollar Auctions: A Multi-Armed Bandit Approach
    Waniek, Marcin
    Long Tran-Tranh
    Michalak, Tomasz
    AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 579 - 587
  • [9] Interface Design Optimization as a Multi-Armed Bandit Problem
    Lomas, J. Derek
    Forlizzi, Jodi
    Poonwala, Nikhil
    Patel, Nirmal
    Shodhan, Sharan
    Patel, Kishan
    Koedinger, Ken
    Brunskill, Emma
    34TH ANNUAL CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2016, 2016, : 4142 - 4153
  • [10] THE MULTI-ARMED BANDIT PROBLEM: AN EFFICIENT NONPARAMETRIC SOLUTION
    Chan, Hock Peng
    ANNALS OF STATISTICS, 2020, 48 (01) : 346 - 373