Mix and Rank: A Framework for Benchmarking Recommender Systems

被引:0
作者
Paudel, Bibek [1 ]
Kocev, Dragi [2 ]
Eftimov, Tome [1 ,2 ]
机构
[1] Stanford Univ, Stanford, CA 94305 USA
[2] Joef Stefan Inst, Ljubljana, Slovenia
来源
2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA) | 2019年
关键词
recommender systems; evaluation; benchmarking; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recommender systems use big data methods, and are widely used in various social-network, e-commerce, and content platforms. With their increased relevance, online platforms and developers are in need of better ways to choose the systems that are most suitable for their use-cases. At the same time, the research literature on recommender systems describes a multitude of measures to evaluate the performance of different algorithms. For the end-user however, the large number of available measures do not provide much help in deciding which algorithm to deploy. Some of the measures are correlated, while others deal with different aspects of recommendation performance like accuracy and coverage. To address this problem, we propose a novel benchmarking framework that mixes different evaluation measures in order to rank the recommender systems on each benchmark dataset, separately. Additionally, our approach discovers sets of correlated measures as well as sets of evaluation measures that are least correlated. We investigate the robustness of the proposed methodology using published results from an experimental study involving multiple big datasets and evaluation measures. Our work provides a general framework that can handle an arbitrary number of evaluation measures and help end-users rank the systems available to them.
引用
收藏
页码:3717 / 3726
页数:10
相关论文
共 50 条
  • [1] The Role of Cores in Recommender Benchmarking for Social Bookmarking Systems
    Doerfel, Stephan
    Jaeschke, Robert
    Stumme, Gerd
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2016, 7 (03)
  • [2] Workshop on Benchmarking Adaptive Retrieval and Recommender Systems - BARS 2013
    Castells, Pablo
    Hopfgartner, Frank
    Said, Alan
    Lalmas, Mounia
    SIGIR'13: THE PROCEEDINGS OF THE 36TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH & DEVELOPMENT IN INFORMATION RETRIEVAL, 2013, : 1133 - 1133
  • [3] BARS: Towards Open Benchmarking for Recommender Systems
    Zhu, Jieming
    Dai, Quanyu
    Su, Liangcai
    Ma, Rong
    Liu, Jinyang
    Cai, Guohao
    Xiao, Xi
    Zhang, Rui
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2912 - 2923
  • [4] Online Learning to Rank for Recommender Systems
    Odijk, Daan
    Schuth, Anne
    PROCEEDINGS OF THE ELEVENTH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'17), 2017, : 348 - 348
  • [5] Introduction to the Special Issue on Recommender System Benchmarking
    Cremonesi, Paolo
    Said, Alan
    Tikk, Domonkos
    Zhou, Michelle X.
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2016, 7 (03)
  • [6] Is Rank Aggregation Effective in Recommender Systems? An Experimental Analysis
    Oliveira, Samuel E. L.
    Diniz, Victor
    Lacerda, Anisio
    Merschmanm, Luiz
    Pappa, Gisele L.
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2020, 11 (02)
  • [7] Rank List Sensitivity of Recommender Systems to Interaction Perturbations
    Oh, Sejoon
    Ustun, Berk
    McAuley, Julian
    Kumar, Srijan
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 1584 - 1594
  • [8] A framework for collaborative filtering recommender systems
    Bobadilla, Jesus
    Hernando, Antonio
    Ortega, Fernando
    Bernal, Jesus
    EXPERT SYSTEMS WITH APPLICATIONS, 2011, 38 (12) : 14609 - 14623
  • [9] Recommender Systems for Privacy Management: A Framework
    Rasmussen, Curtis
    Dara, Rozita
    2014 IEEE 15TH INTERNATIONAL SYMPOSIUM ON HIGH-ASSURANCE SYSTEMS ENGINEERING (HASE), 2014, : 243 - 244
  • [10] ELLIOT: A Comprehensive and Rigorous Framework for Reproducible Recommender Systems Evaluation
    Anelli, Vito Walter
    Bellogin, Alejandro
    Ferrara, Antonio
    Malitesta, Daniele
    Merra, Felice Antonio
    Pomo, Claudio
    Donini, Francesco Maria
    Di Noia, Tommaso
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 2405 - 2414