OTAMatch: Optimal Transport Assignment With PseudoNCE for Semi-Supervised Learning

被引:3
作者
Zhang, Jinjin [1 ]
Liu, Junjie [2 ]
Li, Debang [2 ]
Huang, Qiuyu [2 ]
Chen, Jiaxin [3 ]
Huang, Di [1 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
[2] Meituan, Beijing 100102, Peoples R China
[3] Beihang Univ, Sch Comp Sci & Engn, Lab Intelligent Recognit & Image Proc, Beijing 100191, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Robustness; Predictive models; Noise; Task analysis; Optimization; Data models; Contrastive learning; Pseudo-labeling; semi-supervised learning; optimal transport; contrastive learning; ALLOCATION;
D O I
10.1109/TIP.2024.3425174
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In semi-supervised learning (SSL), many approaches follow the effective self-training paradigm with consistency regularization, utilizing threshold heuristics to alleviate label noise. However, such threshold heuristics lead to the underutilization of crucial discriminative information from the excluded data. In this paper, we present OTAMatch, a novel SSL framework that reformulates pseudo-labeling as an optimal transport (OT) assignment problem and simultaneously exploits data with high confidence to mitigate the confirmation bias. Firstly, OTAMatch models the pseudo-label allocation task as a convex minimization problem, facilitating end-to-end optimization with all pseudo-labels and employing the Sinkhorn-Knopp algorithm for efficient approximation. Meanwhile, we incorporate epsilon-greedy posterior regularization and curriculum bias correction strategies to constrain the distribution of OT assignments, improving the robustness with noisy pseudo-labels. Secondly, we propose PseudoNCE, which explicitly exploits pseudo-label consistency with threshold heuristics to maximize mutual information within self-training, significantly boosting the balance of convergence speed and performance. Consequently, our proposed approach achieves competitive performance on various SSL benchmarks. Specifically, OTAMatch substantially outperforms the previous state-of-the-art SSL algorithms in realistic and challenging scenarios, exemplified by a notable 9.45% error rate reduction over SoftMatch on ImageNet with 100K-label split, underlining its robustness and effectiveness.
引用
收藏
页码:4231 / 4244
页数:14
相关论文
共 73 条
[21]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[22]   Multi-Modal Curriculum Learning for Semi-Supervised Image Classification [J].
Gong, Chen ;
Tao, Dacheng ;
Maybank, Stephen J. ;
Liu, Wei ;
Kang, Guoliang ;
Yang, Jie .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (07) :3249-3260
[23]  
Guo CA, 2017, PR MACH LEARN RES, V70
[24]   Momentum Contrast for Unsupervised Visual Representation Learning [J].
He, Kaiming ;
Fan, Haoqi ;
Wu, Yuxin ;
Xie, Saining ;
Girshick, Ross .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :9726-9735
[25]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[26]  
Hendrycks D., 2016, arXiv:1610.02136
[27]   Label Propagation for Deep Semi-supervised Learning [J].
Iscen, Ahmet ;
Tolias, Giorgos ;
Avrithis, Yannis ;
Chum, Ondrej .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :5065-5074
[28]  
Khosla P, 2020, ADV NEUR IN, V33
[29]  
Kim B., 2021, arXiv
[30]   ConMatch: Semi-supervised Learning with Confidence-Guided Consistency Regularization [J].
Kim, Jiwon ;
Min, Youngjo ;
Kim, Daehwan ;
Lee, Gyuseong ;
Seo, Junyoung ;
Ryoo, Kwangrok ;
Kim, Seungryong .
COMPUTER VISION - ECCV 2022, PT XXX, 2022, 13690 :674-690