SAR Targets Classification Based on Deep Memory Convolution Neural Networks and Transfer Parameters

被引:94
作者
Shang, Ronghua [1 ]
Wang, Jiaming [1 ]
Jiao, Licheng [1 ]
Stolkin, Rustam [2 ]
Hou, Biao [1 ]
Li, Yangyang [1 ]
机构
[1] Xidian Univ, Sch Artificial Intelligence, Key Lab Intelligent Percept & Image Understanding, Minist Educ,Int Res Ctr Intelligent Percept & Com, Xian 710071, Shaanxi, Peoples R China
[2] Univ Birmingham, Extreme Robot Lab, Birmingham B15 2TT, W Midlands, England
基金
英国工程与自然科学研究理事会; 中国国家自然科学基金;
关键词
Deep learning; memory convolutional neural networks (M-Net); parameter transfer; synthetic aperture radar (SAR) targets classification; SPARSE REPRESENTATION; RECOGNITION; IMAGES; PERFORMANCE; ALGORITHM;
D O I
10.1109/JSTARS.2018.2836909
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep learning has obtained state-of-the-art results in a variety of computer vision tasks and has also been used to solve SAR image classification problems. Deep learning algorithms typically require a large amount of training data to achieve high accuracy. In contrast, the size of SAR image datasets is often comparatively limited. Therefore, this paper proposes a novel method, deep memory convolution neural networks (M-Net), to alleviate the problem of overfitting caused by insufficient SAR image samples. Based on the convolutional neural networks (CNN), M-Net adds an information recorder to remember and store samples' spatial features, and then it uses spatial similarity information of the recorded features to predict unknown sample labels. M-Net's use of this information recorder may cause difficulties for convergence if conventional CNN training methods were directly used to train M-Net. To overcome this problem, we propose a transfer parameter technique to train M-Net in two steps. The first step is to train a CNN, which has the same structure as the part of CNN incorporated in M-Net, to obtain initial training parameters. The second step applies the initialized parameters to M-Net and then trains the entire M-Net. This two-step training approach helps us to overcome the nonconvergence issue, and also reduces training time. We evaluate M-Net using the public benchmark MSTAR dataset, and achieve higher accuracy than several other well-known SAR image classification algorithms.
引用
收藏
页码:2834 / 2846
页数:13
相关论文
共 43 条
[1]  
[Anonymous], 2017, P 5 INT C LEARN REPR
[2]  
[Anonymous], PROC CVPR IEEE
[3]  
[Anonymous], 2012, Technical Report
[4]  
[Anonymous], 2015, P 3 INT C LEARN REPR
[5]  
[Anonymous], ACOUST SPEECH SIG PR
[6]  
[Anonymous], 2015, P INT C LEARN REPR I
[7]  
Bengio Y., 2006, ADV NEURAL INFORM PR, V19
[8]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[9]   Large-Scale Machine Learning with Stochastic Gradient Descent [J].
Bottou, Leon .
COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS, 2010, :177-186
[10]   Learning Mid-Level Features For Recognition [J].
Boureau, Y-Lan ;
Bach, Francis ;
LeCun, Yann ;
Ponce, Jean .
2010 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2010, :2559-2566