Distributionally robust learning-to-rank under the Wasserstein metric

被引:0
|
作者
Sotudian, Shahabeddin [1 ]
Chen, Ruidi [1 ]
Paschalidis, Ioannis Ch. [1 ,2 ,3 ]
机构
[1] Boston Univ, Dept Elect & Comp Engn, Div Syst Engn, Boston, MA 02215 USA
[2] Boston Univ, Dept Biomed Engn, Boston, MA 02215 USA
[3] Boston Univ, Fac Comp & Data Sci, Boston, MA 02215 USA
来源
PLOS ONE | 2023年 / 18卷 / 03期
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
ALGORITHM;
D O I
10.1371/journal.pone.0283574
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Despite their satisfactory performance, most existing listwise Learning-To-Rank (LTR) models do not consider the crucial issue of robustness. A data set can be contaminated in various ways, including human error in labeling or annotation, distributional data shift, and malicious adversaries who wish to degrade the algorithm's performance. It has been shown that Distributionally Robust Optimization (DRO) is resilient against various types of noise and perturbations. To fill this gap, we introduce a new listwise LTR model called Distributionally Robust Multi-output Regression Ranking (DRMRR). Different from existing methods, the scoring function of DRMRR was designed as a multivariate mapping from a feature vector to a vector of deviation scores, which captures local context information and cross-document interactions. In this way, we are able to incorporate the LTR metrics into our model. DRMRR uses a Wasserstein DRO framework to minimize a multi-output loss function under the most adverse distributions in the neighborhood of the empirical data distribution defined by a Wasserstein ball. We present a compact and computationally solvable reformulation of the min-max formulation of DRMRR. Our experiments were conducted on two real-world applications: medical document retrieval and drug response prediction, showing that DRMRR notably outperforms state-of-the-art LTR models. We also conducted an extensive analysis to examine the resilience of DRMRR against various types of noise: Gaussian noise, adversarial perturbations, and label poisoning. Accordingly, DRMRR is not only able to achieve significantly better performance than other baselines, but it can maintain a relatively stable performance as more noise is added to the data.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] Distributionally robust day-ahead combined heat and power plants scheduling with Wasserstein Metric
    Skalyga, Mikhail
    Amelin, Mikael
    Wu, Qiuwei
    Soder, Lennart
    ENERGY, 2023, 269
  • [32] A study of distributionally robust mixed-integer programming with Wasserstein metric: on the value of incomplete data
    Ketkov, Sergey S.
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2024, 313 (02) : 602 - 615
  • [33] Wasserstein Distributionally Robust Inverse Multiobjective Optimization
    Dong, Chaosheng
    Zeng, Bo
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 5914 - 5921
  • [34] Wasserstein distributionally robust shortest path problem
    Wang, Zhuolin
    You, Keyou
    Song, Shiji
    Zhang, Yuli
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2020, 284 (01) : 31 - 43
  • [35] Distributionally Robust Stochastic Optimization with Wasserstein Distance
    Gao, Rui
    Kleywegt, Anton
    MATHEMATICS OF OPERATIONS RESEARCH, 2023, 48 (02) : 603 - 655
  • [36] Coresets for Wasserstein Distributionally Robust Optimization Problems
    Huang, Ruomin
    Huang, Jiawei
    Liu, Wenjie
    Ding, Hu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [37] Path Planning Using Wasserstein Distributionally Robust Deep Q-learning
    Alpturk, Cem
    Renganathan, Venkatraman
    2023 EUROPEAN CONTROL CONFERENCE, ECC, 2023,
  • [38] Safe Reinforcement Learning Using Wasserstein Distributionally Robust MPC and Chance Constraint
    Kordabad, Arash Bahari
    Wisniewski, Rafael
    Gros, Sebastien
    IEEE ACCESS, 2022, 10 : 130058 - 130067
  • [39] Computationally Efficient Approximations for Distributionally Robust Optimization Under Moment and Wasserstein Ambiguity
    Cheramin, Meysam
    Cheng, Jianqiang
    Jiang, Ruiwei
    Pan, Kai
    INFORMS JOURNAL ON COMPUTING, 2022, 34 (03) : 1768 - 1794
  • [40] Learning-to-Count by Learning-to-Rank
    D'Alessandro, Adriano C.
    Mahdavi-Amiri, Ali
    Hamarneh, Ghassan
    2023 20TH CONFERENCE ON ROBOTS AND VISION, CRV, 2023, : 105 - 112