Deep Transfer Learning Method Using Self-Pixel and Global Channel Attentive Regularization

被引:1
作者
Kang, Changhee [1 ]
Kang, Sang-ug [1 ]
机构
[1] Sangmyung Univ, Dept Comp Sci, Seoul 03016, South Korea
基金
新加坡国家研究基金会;
关键词
deep transfer learning; knowledge distillation; regularization;
D O I
10.3390/s24113522
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The purpose of this paper is to propose a novel transfer learning regularization method based on knowledge distillation. Recently, transfer learning methods have been used in various fields. However, problems such as knowledge loss still occur during the process of transfer learning to a new target dataset. To solve these problems, there are various regularization methods based on knowledge distillation techniques. In this paper, we propose a transfer learning regularization method based on feature map alignment used in the field of knowledge distillation. The proposed method is composed of two attention-based submodules: self-pixel attention (SPA) and global channel attention (GCA). The self-pixel attention submodule utilizes both the feature maps of the source and target models, so that it provides an opportunity to jointly consider the features of the target and the knowledge of the source. The global channel attention submodule determines the importance of channels through all layers, unlike the existing methods that calculate these only within a single layer. Accordingly, transfer learning regularization is performed by considering both the interior of each single layer and the depth of the entire layer. Consequently, the proposed method using both of these submodules showed overall improved classification accuracy than the existing methods in classification experiments on commonly used datasets.
引用
收藏
页数:11
相关论文
共 37 条
[1]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[2]   Large Scale Fine-Grained Categorization and Domain-Specific Transfer Learning [J].
Cui, Yin ;
Song, Yang ;
Sun, Chen ;
Howard, Andrew ;
Belongie, Serge .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4109-4118
[3]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[4]   Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges [J].
Feng, Di ;
Haase-Schutz, Christian ;
Rosenbaum, Lars ;
Hertlein, Heinz ;
Glaser, Claudius ;
Timm, Fabian ;
Wiesbeck, Werner ;
Dietmayer, Klaus .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (03) :1341-1360
[5]   Borrowing Treasures from the Wealthy: Deep Transfer Learning through Selective Joint Fine-Tuning [J].
Ge, Weifeng ;
Yu, Yizhou .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :10-19
[6]  
Griffin G, 2007, CALTECH 256 OBJECT C
[7]   SpotTune: Transfer Learning through Adaptive Fine-tuning [J].
Guo, Yunhui ;
Shi, Honghui ;
Kumar, Abhishek ;
Grauman, Kristen ;
Rosing, Tajana ;
Feris, Rogerio .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4800-4809
[8]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[9]  
Hinton G, 2015, Arxiv, DOI [arXiv:1503.02531, DOI 10.48550/ARXIV.1503.02531]
[10]   Deep Learning for Emotion Recognition on Small Datasets Using Transfer Learning [J].
Hong-Wei Ng ;
Viet Dung Nguyen ;
Vonikakis, Vassilios ;
Winkler, Stefan .
ICMI'15: PROCEEDINGS OF THE 2015 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2015, :443-449