A survey on unsupervised domain adaptation in computer vision tasks

被引:0
作者
Sun Q. [1 ]
Zhao C. [1 ]
Tang Y. [1 ]
Qian F. [1 ]
机构
[1] School of Information Science and Engineering, East China University of Science and Technology, Shanghai
来源
Zhongguo Kexue Jishu Kexue/Scientia Sinica Technologica | 2022年 / 52卷 / 01期
关键词
Autonomous system; Computer vision; Deep learning; Transfer learning; Unsupervised domain adaptation;
D O I
10.1360/SST-2021-0150
中图分类号
学科分类号
摘要
As one of the typical applications of Industrial Internet, Internet of Vehicles (IoV) develops rapidly in recent years. It relies on the interconnection of information, where the transferability of the accurate perception is of great importance. Though deep learning has accelerated the development of computer vision tasks, traditional deep learning-based methods still have strong reliance on manually annotated training data and are poor at generalizing knowledge to new environments. For computer vision tasks, since it is difficult to collect training data with ground truth, it is agent to improve the generalization ability of deep learning models and alleviate their dependence on manually annotated labels. Unsupervised domain adaptation (UDA) methods apply deep learning models to extract and align features from data in different domains, which ensures the satisfactory generalization performance of deep learning-based computer vision algorithms. This paper focuses on the challenges and applications of UDA in some typical computer vision tasks. Firstly, the definition, significances, application difficulties, basic methods and relevant datasets of deep learning-based UDA methods are introduced. Then, the mainstream UDA methods in typical computer vision tasks are introduced separately. Finally, the prospective technical development trends and a summary are given. © 2022, Science Press. All right reserved.
引用
收藏
页码:26 / 54
页数:28
相关论文
共 174 条
[1]  
Yuan Y, Tang X, Zhou W, Et al., Data driven discovery of cyber physical systems, Nature Commun, 10, pp. 1-9, (2019)
[2]  
Yuan Y, Ma G, Cheng C, Et al., A general end-to-end diagnosis framework for manufacturing systems, Natl Sci Rev, 7, pp. 418-429, (2020)
[3]  
Tang Y, Zhao C, Wang J, Et al., Perception and decision-making of autonomous systems in the era of learning: An overview, (2020)
[4]  
Li H, Duan H B., Verification of monocular and binocular pose estimation algorithms in vision-based uavs autonomous aerial refueling system, Sci China Tech Sci, 59, pp. 1730-1738, (2016)
[5]  
Zhu Z S, Su A, Liu H B, Et al., Vision navigation for aircrafts based on 3D reconstruction from real-time image sequences, Sci China Tech Sci, 58, pp. 1196-1208, (2015)
[6]  
Krizhevsky A, Sutskever I, Hinton G E., Imagenet classification with deep convolutional neural networks, Proceedings of the 2012 Advances in Neural Information Processing Systems (NIPS), pp. 1097-1105, (2012)
[7]  
He K, Zhang X, Ren S, Et al., Deep residual learning for image recognition, Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770-778, (2016)
[8]  
Ren S, He K, Girshick R, Et al., Faster r-cnn: Towards real-time object detection with region proposal networks, Proceedings of the 2015 Advances in Neural Information Processing Systems (NIPS), pp. 91-99, (2015)
[9]  
Long J, Shelhamer E, Darrell T., Fully convolutional networks for semantic segmentation, Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3431-3440, (2015)
[10]  
Chen L C, Papandreou G, Kokkinos I, Et al., Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs, IEEE Trans Pattern Anal Mach Intell, 40, pp. 834-848, (2017)