Decorrelating structure via adapters makes ensemble learning practical for Semi-supervised Learning

被引:0
作者
Wu, Jiaqi [1 ]
Pang, Junbiao [1 ]
Huang, Qingming [2 ]
机构
[1] Beijing University of Technology, College of Computer Science, No. 100, Pingleyuan, Chaoyang District, Beijing
[2] University of Chinese Academy of Sciences, College of Computer Science and Engineering, No. 80, Zhongguancun East Road, Haidian District, Beijing
关键词
Classification; Ensemble learning; Pose estimation; Semi-supervised learning;
D O I
10.1016/j.engappai.2025.111641
中图分类号
学科分类号
摘要
In Semi-Supervised Learning (SSL), ensemble learning boosts training by enhancing pseudo-label accuracy through multi-perspective prediction integration. This improvement hinges on ensemble framework diversity. However, balancing ensemble cost and diversity is crucial. High costs can impede practicality or efficiency, while low diversity limits performance gains. Our experiments comparing ensemble methods reveal that diversity is influenced by prediction similarity among predictors. Thus, we propose ”Decorrelating Structure via Adapters” (DSA), using distinct adapters to map input features of each prediction head into independent spaces, fostering diversity. As a lightweight structure, DSA effectively decorrelates without substantially increasing parameters or relying on extra loss functions. We analyze DSA's mechanism, explaining how adapters reduce prediction similarity. Moreover, we showcase DSA's significant improvements in semi-supervised classification and pose estimation tasks. The code is publicly accessible at: https://github.com/Qi2019KB/DSA. © 2025 Elsevier Ltd
引用
收藏
相关论文
共 48 条
[1]  
Bachman P., Alsharif O., Precup D., Learning with pseudo-ensembles, Adv. Neural Inf. Process. Syst., 27, (2014)
[2]  
Breiman L., Bagging predictors, Mach. Learn., 24, pp. 123-140, (1996)
[3]  
Brown G., Diversity in Neural Network Ensembles, (2004)
[4]  
Cubuk E.D., Zoph B., Shlens J., pp. 702-703, (2020)
[5]  
Dietterich T.G., Ensemble methods in machine learning, International Workshop on Multiple Classifier Systems, pp. 1-15, (2000)
[6]  
Han B., Yao Q., Yu X., Niu G., Xu M., Hu W., Tsang I., Sugiyama M., Co-teaching: Robust training of deep neural networks with extremely noisy labels, Adv. Neural Inf. Process. Syst., 31, (2018)
[7]  
Hendrycks D., Dietterich T., Benchmarking neural network robustness to common corruptions and perturbations, (2019)
[8]  
Ho T.K., The random subspace method for constructing decision forests, IEEE Trans. Pattern Anal. Mach. Intell., 20, 8, pp. 832-844, (1998)
[9]  
Huang G., Li Y., Pleiss G., Liu Z., Hopcroft J.E., Weinberger K.Q., Snapshot ensembles: Train 1, get m for free, (2017)
[10]  
Huang Z., Shen L., Yu J., Han B., Liu T., Flatmatch: Bridging labeled data and unlabeled data with cross-sharpness for semi-supervised learning, Adv. Neural Inf. Process. Syst., 36, pp. 18474-18494, (2023)