A genetic algorithm-based virtual sample generation technique to improve small data set learning

被引:88
作者
Li, Der-Chiang [1 ]
Wen, I-Hsiang [1 ]
机构
[1] Natl Cheng Kung Univ, Dept Ind & Informat Management, Tainan 70101, Taiwan
关键词
Small data set; Genetic algorithm-based virtual sample generation (GABVSG); Feasibility-based programming (FBP) model; NEURAL-NETWORK; TREND-DIFFUSION; KNOWLEDGE; MODEL; SIZE;
D O I
10.1016/j.neucom.2014.06.004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While back-propagation neural networks (BPNN) are effective learning tools for building non-linear models, they are often unstable when using small-data-sets. Therefore, in order to solve this problem, we construct artificial samples, called virtual samples, to improve the learning robustness. This research develops a novel method of virtual sample generation (VSG), named genetic algorithm-based virtual sample generation (GABVSG), which considers the integrated effects and constraints of data attributes. We first determine the acceptable range by using mega-trend diffusion (MTD) functions, and construct the feasibility-based programming (FBP) model with BPNN. A genetic algorithm (GA) is then applied to accelerate the generation of the most-feasible virtual samples. Finally, we use two real cases to verify the performance of the proposed method by comparing the results with two well-known forecasting models, BPNN, support vector machine for regression (SVR) and one newly published approach MTD method [1]. The experimental results indicate that the performance of the GABVSG method is superior to that of using original training data without artificial samples. Consequently, the proposed method can improve learning performance significantly when working with small samples. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:222 / 230
页数:9
相关论文
共 32 条
[1]   A meta-learning approach to automatic kernel selection for support vector machines [J].
Ali, Shawkat ;
Smith-Miles, Kate A. .
NEUROCOMPUTING, 2006, 70 (1-3) :173-186
[2]   Improving support vector machine classifiers by modifying kernel functions [J].
Amari, S ;
Wu, S .
NEURAL NETWORKS, 1999, 12 (06) :783-789
[3]   Binarized Support Vector Machines [J].
Carrizosa, Emilio ;
Martin-Barragan, Belen ;
Morales, Dolores Romero .
INFORMS JOURNAL ON COMPUTING, 2010, 22 (01) :154-167
[4]   A new approach to prediction of radiotherapy of bladder cancer cells in small dataset analysis [J].
Chao, Gy-Yi ;
Tsai, Tung-I ;
Lu, Te-Jung ;
Hsu, Hung-Chang ;
Bao, Bo-Ying ;
Wu, Wan-Yu ;
Lin, Miao-Ting ;
Lu, Te-Ling .
EXPERT SYSTEMS WITH APPLICATIONS, 2011, 38 (07) :7963-7969
[5]  
Cherkassky V, 1997, IEEE Trans Neural Netw, V8, P1564, DOI 10.1109/TNN.1997.641482
[6]   A SENSITIVITY ANALYSIS OF A BACK-PROPAGATION NEURAL NETWORK FOR MANUFACTURING PROCESS PARAMETERS [J].
COOK, DF ;
SHANNON, RE .
JOURNAL OF INTELLIGENT MANUFACTURING, 1991, 2 (03) :155-163
[7]   Retrieval parameter optimization using genetic algorithms [J].
Fujita, Sumio .
INFORMATION PROCESSING & MANAGEMENT, 2009, 45 (06) :664-682
[8]  
Golberg D. E., 1989, GENETIC ALGORITHMS S, V1989, P36
[9]  
Holland J.H., 1992, Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control and Artificial Intelligence
[10]   A diffusion-neural-network for learning from small samples [J].
Huang, CF ;
Moraga, C .
INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2004, 35 (02) :137-161