Efficient and Effective Tree-based and Neural Learning to Rank

被引:11
作者
Bruch, Sebastian [1 ]
Lucchese, Claudio [2 ]
Nardini, Franco Maria [3 ]
机构
[1] Pinecone, San Francisco, CA 94104 USA
[2] Ca Foscari Univ, Venice, Italy
[3] CNR, ISTI, Pisa, Italy
来源
FOUNDATIONS AND TRENDS IN INFORMATION RETRIEVAL | 2023年 / 17卷 / 01期
关键词
INFORMATION-RETRIEVAL; TRAVERSAL; ENSEMBLES;
D O I
10.1561/1500000071
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As information retrieval researchers, we not only develop algorithmic solutions to hard problems, but we also insist on a proper, multifaceted evaluation of ideas. The literature on the fundamental topic of retrieval and ranking, for instance, has a rich history of studying the effectiveness of indexes, retrieval algorithms, and complex machine learning rankers, while at the same time quantifying their computational costs, from creation and training to application and inference. This is evidenced, for example, by more than a decade of research on efficient training and inference of large decision forest models in Learning to Rank (LtR). As we move towards even more complex, deep learning models in a wide range of applications, questions on efficiency have once again resurfaced with renewed urgency. Indeed, efficiency is no longer limited to time and space; instead it has found new, challenging dimensions that stretch to resource-, sample- and energy-efficiency with ramifications for researchers, users, and the environment. This monograph takes a step towards promoting the study of efficiency in the era of neural information retrieval by offering a comprehensive survey of the literature on efficiency and effectiveness in ranking, and to a limited extent, retrieval. This monograph was inspired by the parallels that exist between the challenges in neural network-based ranking solutions and their predecessors, decision forest-based LtR models, as well as the connections between the solutions the literature to date has to offer. We believe that by understanding the fundamentals underpinning these algorithmic and data structure solutions for containing the contentious relationship between efficiency and effectiveness, one can better identify future directions and more efficiently determine the merits of ideas. We also present what we believe to be important research directions in the forefront of efficiency and effectiveness in retrieval and ranking.
引用
收藏
页码:1 / 123
页数:123
相关论文
共 216 条
[41]   Out-of-Domain Semantics to the Rescue! Zero-Shot Hybrid Retrieval Models [J].
Chen, Tao ;
Zhang, Mingyang ;
Lu, Jing ;
Bendersky, Michael ;
Najork, Marc .
ADVANCES IN INFORMATION RETRIEVAL, PT I, 2022, 13185 :95-110
[42]   XGBoost: A Scalable Tree Boosting System [J].
Chen, Tianqi ;
Guestrin, Carlos .
KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, :785-794
[43]  
Chuklin Aleksandr, 2015, Click Models for Web Search
[44]   Universal Approximation Functions for Fast Learning to Rank [J].
Cohen, Daniel ;
Foley, John ;
Zamani, Hamed ;
Allan, James ;
Croft, W. Bruce .
ACM/SIGIR PROCEEDINGS 2018, 2018, :1017-1020
[45]   Efficient and effective spam filtering and re-ranking for large web datasets [J].
Cormack, Gordon V. ;
Smucker, Mark D. ;
Clarke, Charles L. A. .
INFORMATION RETRIEVAL, 2011, 14 (05) :441-465
[46]  
Culpepper J. Shane, 2016, P 21 AUSTRALASIAN DO, P17
[47]  
Cuturi M., 2019, Advances in Neural Information Processing Systems, V32
[48]   Deeper Text Understanding for IR with Contextual Neural Language Modeling [J].
Dai, Zhuyun ;
Callan, Jamie .
PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, :985-988
[49]   Convolutional Neural Networks for Soft-Matching N-Grams in Ad-hoc Search [J].
Dai, Zhuyun ;
Xiong, Chenyan ;
Callan, Jamie ;
Liu, Zhiyuan .
WSDM'18: PROCEEDINGS OF THE ELEVENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2018, :126-134
[50]  
Dang Van, 2013, Proceedings, P423