SELFOOD: Self-Supervised Out-Of-Distribution Detection via Learning to Rank

被引:0
作者
Mekalas, Dheeraj [1 ]
Samavedhi, Adithya [1 ]
Dong, Chengyu [1 ]
Shang, Jingbo [1 ,2 ]
机构
[1] Univ Calif San Diego, La Jolla, CA 92093 USA
[2] Univ Calif San Diego, Halicioglu Data Sci Inst, La Jolla, CA 92093 USA
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023) | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural classifiers trained with crossentropy (CE) loss often suffer from poor calibration, necessitating the task of out-ofdistribution (OOD) detection. Traditional supervised OOD detection methods require expensive manual annotation of in-distribution and OOD samples. To address the annotation bottleneck, we introduce SELFOOD, a selfsupervised OOD detection method that requires only in-distribution samples as supervision. We cast OOD detection as an inter-document intralabel (IDIL) ranking problem and train the classifier with our pairwise ranking loss, referred to as IDIL loss. Specifically, given a set of in-distribution documents and their labels, for each label, we train the classifier to rank the softmax scores of documents belonging to that label to be higher than the scores of documents that belong to other labels. Unlike CE loss, our IDIL loss reaches zero when the desired confidence ranking is achieved and gradients are backpropagated to decrease probabilities associated with incorrect labels rather than continuously increasing the probability of the correct label. Extensive experiments with several classifiers on multiple classification datasets demonstrate the effectiveness of our method in both coarse-and fine-grained settings.
引用
收藏
页码:10721 / 10734
页数:14
相关论文
共 46 条
[1]  
Nguyen A, 2015, PROC CVPR IEEE, P427, DOI 10.1109/CVPR.2015.7298640
[2]  
[Anonymous], 2022, P 29 INT C COMP LING
[3]  
[Anonymous], 2019, CoRR, DOI DOI 10.48550/arXiv.1907.11692
[4]  
Brown TB, 2020, ADV NEUR IN, V33
[5]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[6]  
DeVries T, 2018, ARXIV
[7]   Sigmoid-weighted linear units for neural network function approximation in reinforcement learning [J].
Elfwing, Stefan ;
Uchibe, Eiji ;
Doya, Kenji .
NEURAL NETWORKS, 2018, 107 :3-11
[8]  
Fei G., 2016, P 2016 C N AM CHAPT, P506, DOI DOI 10.18653/V1/N16-1061
[9]  
Gangal V, 2020, AAAI CONF ARTIF INTE, V34, P7764
[10]  
GOOD IJ, 1952, J ROY STAT SOC B, V14, P107