Fusion based Heterogeneous Convolutional Neural Networks Architecture

被引:0
作者
Komish, David [1 ]
Ezekiel, Soundararajan [1 ]
Comacchia, Maria [2 ]
机构
[1] Indiana Univ Penn, Indiana, PA 15705 USA
[2] US Air Force, Res Lab, Rome, NY USA
来源
2018 IEEE APPLIED IMAGERY PATTERN RECOGNITION WORKSHOP (AIPR) | 2018年
关键词
Convolutional Neural Network; Support Vector Machine; Classifier; Principal component analysis; heterogeneous network; RECOGNITION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, Deep Convolutional Neural Networks (DCNNs) have gained lots of attention and won many competitions in machine learning, object detection, image classification, and pattern recognition. The breakthroughs in the development of graphical processing units have made it possible to train DCNNs quickly for state-of-the-art tasks such as image classification, speech recognition, and many others. However, to solve complex problems, these multilayered convolutional neural networks become increasingly large, complex, and abstract. We propose methods to improve the performance of neural networks while reducing their dimensionality, enabling a better understanding of the learning process. To leverage the extensive training, as well as strengths of several pretrained models, we explored new approaches for combining features from fully connected layers of models with heterogeneous architectures. The proposed approach combines features extracted from the penultimate fully connected layer from three different DCNNs. We merge the features of all three DCNNs together and apply principal component analysis or linear discriminant analysis. Our approach aims to reduce the dimensionality of the feature vector and find the smallest feature vector dimension that can maintain the classifier performance. For this task we use a linear Support Vector Machine as a classifier. We also investigate whether it is advantageous to fuse only penultimate fully connected layers, or to perform fusion based on other fully connected layers using multiple homogenous or heterogeneous networks. The results show that the fusion method outperformed both individual networks in terms of accuracy and computational time in all of our various trial sizes. Overall our fusion methods are faster and more accurate than individual networks in both training and testing. Finally, we compared heterogeneous with homogenous fusion methods and the results show heterogeneous methods outperform homogeneous methods.
引用
收藏
页数:6
相关论文
共 22 条
[1]  
Abdi Herve, 2010, PRINCIPLE COMPONENT, V2
[2]  
[Anonymous], 2017, TRANSF LEARN US AL
[3]  
[Anonymous], 1989, Neural Information Processing Systems
[4]  
[Anonymous], 2017, QUARTZ
[5]  
[Anonymous], J MACHINE LEARNING R
[6]  
[Anonymous], ICLR
[7]   A tutorial on Support Vector Machines for pattern recognition [J].
Burges, CJC .
DATA MINING AND KNOWLEDGE DISCOVERY, 1998, 2 (02) :121-167
[8]  
Deng J., 2009, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2009.5206848
[10]  
Guyon Isabelle, 2005, INTRO FEATURE EXTRAC