Efficient Ensemble via Rotation-Based Self- Supervised Learning Technique and Multi-Input Multi-Output Network

被引:1
作者
Park, Jaehoon [1 ]
机构
[1] Korea Univ, Dept Artificial Intelligence, Seoul 02841, South Korea
关键词
MIMO communication; Self-supervised learning; Task analysis; Feature extraction; Vectors; Training; Periodic structures; Convolutional neural networks; Deep learning; Ensemble learning; Convolution neural networks (CNNs); deep ensemble; multi-input multi-output network;
D O I
10.1109/ACCESS.2024.3373692
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-input multi-output structures have been developed to boost performance by learning multiple ensemble members at a small additional cost to a single network. There were several attempts to further develop multi-input multi-output structures; however, integrating the benefits of self-supervised learning into a multi-input multi-output structure has not yet been studied. In this work, we develop a multi-input multi-output structure designed to jointly learn original and self-supervised tasks, thereby leveraging the benefits of self-supervised learning. Specifically, in terms of multiple inputs, we improve the mixing strategy and minibatch structure for rotation-based self-supervised learning technique, and in terms of multiple outputs, we extend the label space of multiple classifiers to predict both the original class and true rotation degree. We observe that our method with wider networks on CIFAR-10, CIFAR-100, and Tiny ImageNet datasets shows better performance compared to previous works, even with nearly half the number of parameters, e.g., using only about 45.8% of the number of parameters compared to the best-performing multi-input multi-output method, MixMo, in the Tiny ImageNet dataset, while still achieving a 2.01% improvement.
引用
收藏
页码:36135 / 36147
页数:13
相关论文
共 35 条
[1]  
[Anonymous], 1999, Proc. J. Artif. Intell. Res., V11, P169
[2]  
Barber D., 1998, Neural Networks and Machine Learning. Proceedings, P215
[3]  
Chrabaszcz P, 2017, Arxiv, DOI [arXiv:1707.08819, DOI 10.48550/ARXIV.1707.08819]
[4]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[5]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[6]   Unsupervised Visual Representation Learning by Context Prediction [J].
Doersch, Carl ;
Gupta, Abhinav ;
Efros, Alexei A. .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1422-1430
[7]  
Fort S, 2020, Arxiv, DOI [arXiv:1912.02757, 10.48550/arXiv.1912.02757]
[8]  
Gal Y, 2016, PR MACH LEARN RES, V48
[9]   Ensemble deep learning: A review [J].
Ganaie, M. A. ;
Hu, Minghui ;
Malik, A. K. ;
Tanveer, M. ;
Suganthan, P. N. .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 115
[10]  
Gidaris P., 2018, ICLR