OTAMatch: Optimal Transport Assignment With PseudoNCE for Semi-Supervised Learning

被引:3
作者
Zhang, Jinjin [1 ]
Liu, Junjie [2 ]
Li, Debang [2 ]
Huang, Qiuyu [2 ]
Chen, Jiaxin [3 ]
Huang, Di [1 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
[2] Meituan, Beijing 100102, Peoples R China
[3] Beihang Univ, Sch Comp Sci & Engn, Lab Intelligent Recognit & Image Proc, Beijing 100191, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Robustness; Predictive models; Noise; Task analysis; Optimization; Data models; Contrastive learning; Pseudo-labeling; semi-supervised learning; optimal transport; contrastive learning; ALLOCATION;
D O I
10.1109/TIP.2024.3425174
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In semi-supervised learning (SSL), many approaches follow the effective self-training paradigm with consistency regularization, utilizing threshold heuristics to alleviate label noise. However, such threshold heuristics lead to the underutilization of crucial discriminative information from the excluded data. In this paper, we present OTAMatch, a novel SSL framework that reformulates pseudo-labeling as an optimal transport (OT) assignment problem and simultaneously exploits data with high confidence to mitigate the confirmation bias. Firstly, OTAMatch models the pseudo-label allocation task as a convex minimization problem, facilitating end-to-end optimization with all pseudo-labels and employing the Sinkhorn-Knopp algorithm for efficient approximation. Meanwhile, we incorporate epsilon-greedy posterior regularization and curriculum bias correction strategies to constrain the distribution of OT assignments, improving the robustness with noisy pseudo-labels. Secondly, we propose PseudoNCE, which explicitly exploits pseudo-label consistency with threshold heuristics to maximize mutual information within self-training, significantly boosting the balance of convergence speed and performance. Consequently, our proposed approach achieves competitive performance on various SSL benchmarks. Specifically, OTAMatch substantially outperforms the previous state-of-the-art SSL algorithms in realistic and challenging scenarios, exemplified by a notable 9.45% error rate reduction over SoftMatch on ImageNet with 100K-label split, underlining its robustness and effectiveness.
引用
收藏
页码:4231 / 4244
页数:14
相关论文
共 73 条
[1]  
Altschuler J., 2017, Adv. Neural Inf. Process. Syst., V30, P1
[2]  
[Anonymous], 2013, JMLR
[3]  
[Anonymous], 2004, P ADV NEUR INF PROC
[4]   Pseudo-Labeling and Confirmation Bias in Deep Semi-Supervised Learning [J].
Arazo, Eric ;
Ortego, Diego ;
Albert, Paul ;
O'Connor, Noel E. ;
McGuinness, Kevin .
2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
[5]  
Asano Y.M., 2020, ICLR
[6]  
Bengio Y., 2009, P 26 ANN INT C MACH, P41
[7]  
Berthelot D., 2020, INT C LEARN REPR ICL
[8]  
Berthelot D, 2019, ADV NEUR IN, V32
[9]   End-to-End Object Detection with Transformers [J].
Carion, Nicolas ;
Massa, Francisco ;
Synnaeve, Gabriel ;
Usunier, Nicolas ;
Kirillov, Alexander ;
Zagoruyko, Sergey .
COMPUTER VISION - ECCV 2020, PT I, 2020, 12346 :213-229
[10]  
Caron M, 2020, ADV NEUR IN, V33