Representation Similarity Analysis for Efficient Task Taxonomy & Transfer Learning

被引:62
作者
Dwivedi, Kshitij [1 ]
Roig, Gemma [1 ]
机构
[1] Singapore Univ Technol & Design, Singapore, Singapore
来源
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019) | 2019年
关键词
D O I
10.1109/CVPR.2019.01267
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transfer learning is widely used in deep neural network models when there are few labeled examples available. The common approach is to take a pre-trained network in a similar task and finetune the model parameters. This is usually done blindly without a pre-selection from a set of pre-trained models, or by finetuning a set of models trained on different tasks and selecting the best performing one by cross-validation. We address this problem by proposing an approach to assess the relationship between visual tasks and their task-specific models. Our method uses Representation Similarity Analysis (RSA), which is commonly used to find a correlation between neuronal responses from brain data and models. With RSA we obtain a similarity score among tasks by computing correlations between models trained on different tasks. Our method is efficient as it requires only pre-trained models, and a few images with no further training. We demonstrate the effectiveness and efficiency of our method to generating task taxonomy on Taskonomy dataset. We next evaluate the relationship of RSA with the transfer learning performance on Taskonomy tasks and a new task: Pascal VOC semantic segmentation. Our results reveal that models trained on tasks with higher similarity score show higher transfer learning performance. Surprisingly, the best transfer learning result for Pascal VOC semantic segmentation is not obtained from the pre-trained model on semantic segmentation, probably due to the domain differences, and our method successfully selects the high performing models.
引用
收藏
页码:12379 / 12388
页数:10
相关论文
共 35 条
[1]  
Bonner Michael F, PLOS COMPUTATIONAL B
[2]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[3]  
Chen LB, 2017, IEEE INT SYMP NANO, P1, DOI 10.1109/NANOARCH.2017.8053709
[4]   Dynamics of scene representations in the human brain revealed by magnetoencephalography and deep neural networks [J].
Cichy, Radoslaw Martin ;
Khosla, Aditya ;
Pantazis, Dimitrios ;
Oliva, Aude .
NEUROIMAGE, 2017, 153 :346-358
[5]   Comparison of deep neural networks to spatio-temporal cortical dynamics of human visual object recognition reveals hierarchical correspondence [J].
Cichy, Radoslaw Martin ;
Khosla, Aditya ;
Pantazis, Dimitrios ;
Torralba, Antonio ;
Oliva, Aude .
SCIENTIFIC REPORTS, 2016, 6
[6]   Resolving human object recognition in space and time [J].
Cichy, Radoslaw Martin ;
Pantazis, Dimitrios ;
Oliva, Aude .
NATURE NEUROSCIENCE, 2014, 17 (03) :455-462
[7]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[8]  
Dharmasiri T, 2017, IEEE INT C INT ROBOT, P1505, DOI 10.1109/IROS.2017.8205954
[9]  
Dvornik Nikita, BLITZNET REAL TIME D
[10]   The Pascal Visual Object Classes (VOC) Challenge [J].
Everingham, Mark ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) :303-338