Tune and mix: learning to rank using ensembles of calibrated multi-class classifiers

被引:0
作者
Róbert Busa-Fekete
Balázs Kégl
Tamás Éltető
György Szarvas
机构
[1] University of Paris-Sud,Linear Accelerator Laboratory (LAL)
[2] CNRS,Linear Accelerator Laboratory (LAL) and Computer Science Laboratory (LRI)
[3] Research Group on Artificial Intelligence (RGAI) of the Hungarian Academy of Sciences and University of Szeged,undefined
[4] University of Paris-Sud,undefined
[5] CNRS,undefined
[6] Ericsson Hungary,undefined
[7] Nuance Communications Germany GmbH,undefined
来源
Machine Learning | 2013年 / 93卷
关键词
Learning-to-rank; Multi-class classification; Class Probability Calibration; Regression Based Calibration; Ensemble methods;
D O I
暂无
中图分类号
学科分类号
摘要
In subset ranking, the goal is to learn a ranking function that approximates a gold standard partial ordering of a set of objects (in our case, a set of documents retrieved for the same query). The partial ordering is given by relevance labels representing the relevance of documents with respect to the query on an absolute scale. Our approach consists of three simple steps. First, we train standard multi-class classifiers (AdaBoost.MH and multi-class SVM) to discriminate between the relevance labels. Second, the posteriors of multi-class classifiers are calibrated using probabilistic and regression losses in order to estimate the Bayes-scoring function which optimizes the Normalized Discounted Cumulative Gain (NDCG). In the third step, instead of selecting the best multi-class hyperparameters and the best calibration, we mix all the learned models in a simple ensemble scheme.
引用
收藏
页码:261 / 292
页数:31
相关论文
共 47 条
[1]  
Audibert J. Y.(2009)Exploration-exploitation tradeoff using variance estimates in multi-armed bandits Theoretical Computer Science 410 1876-1902
[2]  
Munos R.(2002)The non-stochastic multi-armed bandit problem SIAM Journal on Computing 32 48-77
[3]  
Szepesvári C.(2012)MultiBoost: a multi-purpose boosting package Journal of Machine Learning Research 13 549-553
[4]  
Auer P.(2010)Gradient descent optimization of smoothed information retrieval metrics Information Retrievel 13 216-235
[5]  
Cesa-Bianchi N.(2008)Statistical analysis of Bayes optimal subset ranking IEEE Transactions on Information Theory 54 5140-5154
[6]  
Freund Y.(2001)On the algorithmic implementation of multiclass kernel-based vector machines Journal of Machine Learning Research 2 265-292
[7]  
Schapire R.(1997)A decision-theoretic generalization of on-line learning and an application to boosting Journal of Computer and System Sciences 55 119-139
[8]  
Benbouzid D.(2003)An efficient boosting algorithm for combining preferences Journal of Machine Learning Research 4 933-969
[9]  
Busa-Fekete R.(2002)Stochastic gradient boosting Computational Statistics & Data Analysis 38 367-378
[10]  
Casagrande N.(2002)Cumulated gain-based evaluation of IR techniques ACM Transactions on Information Systems 20 422-446