Deep Transfer Learning Method Using Self-Pixel and Global Channel Attentive Regularization

被引:1
作者
Kang, Changhee [1 ]
Kang, Sang-ug [1 ]
机构
[1] Sangmyung Univ, Dept Comp Sci, Seoul 03016, South Korea
基金
新加坡国家研究基金会;
关键词
deep transfer learning; knowledge distillation; regularization;
D O I
10.3390/s24113522
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The purpose of this paper is to propose a novel transfer learning regularization method based on knowledge distillation. Recently, transfer learning methods have been used in various fields. However, problems such as knowledge loss still occur during the process of transfer learning to a new target dataset. To solve these problems, there are various regularization methods based on knowledge distillation techniques. In this paper, we propose a transfer learning regularization method based on feature map alignment used in the field of knowledge distillation. The proposed method is composed of two attention-based submodules: self-pixel attention (SPA) and global channel attention (GCA). The self-pixel attention submodule utilizes both the feature maps of the source and target models, so that it provides an opportunity to jointly consider the features of the target and the knowledge of the source. The global channel attention submodule determines the importance of channels through all layers, unlike the existing methods that calculate these only within a single layer. Accordingly, transfer learning regularization is performed by considering both the interior of each single layer and the depth of the entire layer. Consequently, the proposed method using both of these submodules showed overall improved classification accuracy than the existing methods in classification experiments on commonly used datasets.
引用
收藏
页数:11
相关论文
共 37 条
[11]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]
[12]   Lung and Pancreatic Tumor Characterization in the Deep Learning Era: Novel Supervised and Unsupervised Learning Approaches [J].
Hussein, Sarfaraz ;
Kandel, Pujan ;
Bolan, Candice W. ;
Wallace, Michael B. ;
Bagci, Ulas .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2019, 38 (08) :1777-1787
[13]   Self-Supervised Denoising Image Filter Based on Recursive Deep Neural Network Structure [J].
Kang, Changhee ;
Kang, Sang-ug .
SENSORS, 2021, 21 (23)
[14]  
Khosla A., 2011, 1 WORKSH FIN GRAIN V
[15]   Few Sample Knowledge Distillation for Efficient Network Compression [J].
Li, Tianhong ;
Li, Jianguo ;
Liu, Zhuang ;
Zhang, Changshui .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :14627-14635
[16]  
Li XJ, 2020, Arxiv, DOI [arXiv:1901.09229, DOI 10.48550/ARXIV.1901.09229]
[17]   A baseline regularization scheme for transfer learning with convolutional neural networks [J].
Li, Xuhong ;
Grandvalet, Yves ;
Davoine, Franck .
PATTERN RECOGNITION, 2020, 98
[18]  
Li Xuhong, 2018, P MACHINE LEARNING R, V80
[19]   Learning without Forgetting [J].
Li, Zhizhong ;
Hoiem, Derek .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (12) :2935-2947
[20]  
Mirzadeh SI, 2020, AAAI CONF ARTIF INTE, V34, P5191