NORMALIZED DATA BARRIER AMPLIFIER FOR FEED-FORWARD NEURAL NETWORK

被引:1
作者
Fuangkhon, P. [1 ]
机构
[1] Assumption Univ, Dept Digital Business Management, 88 Moo 8 Bangna Trad KM 26, Bang Sao Thong 10540, Samut Prakan, Thailand
关键词
data mining; data normalization; data reduction; instance selection; neural network;
D O I
10.14311/NNW.2021.31.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A boundary vector generator is a data barrier amplifier that improves the distribution model of the samples to increase the classification accuracy of the feed-forward neural network. It generates new forms of samples, one for amplifying the barrier of their class (fundamental multi-class outpost vectors) and the other for increasing the barrier of the nearest class (additional multi-class outpost vectors). However, these sets of boundary vectors are enormous. The reduced boundary vector generators proposed three boundary vector reduction techniques that scale down fundamental multi-class outpost vectors and additional multi-class outpost vectors. Nevertheless, these techniques do not consider the interval of the attributes, causing some attributes to suppress over the other attributes on the Euclidean distance calculation. The motivation of this study is to explore whether six normalization techniques; min-max, Z-score, mean and mean absolute deviation, median and median absolute deviation, modified hyperbolic tangent, and hyperbolic tangent estimator, can improve the classification performance of the boundary vector generator and the reduced boundary vector generators for maximizing class boundary. Each normalization technique pre-processes the original training set before the boundary vector generator or each of the three reduced boundary vector generators will begin. The experimental results on the real-world datasets generally confirmed that (1) the final training set having only FF-AA reduced boundary vectors can be integrated with one of the normalization techniques effectively when the accuracy and precision are prioritized, (2) the final training set having only the boundary vectors can be integrated with one of the normalization techniques effectively when the recall and F1-score are prioritized, (3) the Z-score normalization can generally improve the accuracy and precision of all types of training sets, (4) the modified hyperbolic tangent normalization can generally improve the recall of all types of training sets, (5) the min-max normalization can generally improve the accuracy and F1-score of all types of training sets, and (6) the selection of the normalization techniques and the training set types depends on the key performance measure for the dataset.
引用
收藏
页码:125 / 157
页数:33
相关论文
共 17 条
[1]  
[Anonymous], 2011, International Journal of Computer Applications
[2]  
[Anonymous], 2015, 135282015 ISO, DOI [10.31030/3116750, DOI 10.31030/3116750]
[3]  
[Anonymous], 2017, UCI Machine Learning Repository
[4]  
Burgeon C, 2017, DOMITIEN EMPEREUR CO
[5]  
CORTES C, 1995, MACH LEARN, V20, P273, DOI 10.1023/A:1022627411411
[6]   NEAREST NEIGHBOR PATTERN CLASSIFICATION [J].
COVER, TM ;
HART, PE .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1967, 13 (01) :21-+
[7]  
Fuangkhon Piyabute, 2012, Intelligent Data Engineering and Automated Learning - IDEAL 2012. Proceedings 13th International Conference, P35, DOI 10.1007/978-3-642-32639-4_5
[8]   Reduced Multi-class Contour Preserving Classification [J].
Fuangkhon, Piyabute ;
Tanprasert, Thitipong .
NEURAL PROCESSING LETTERS, 2016, 43 (03) :759-804
[9]  
Han J, 2012, MOR KAUF D, P1
[10]   Score normalization in multimodal biometric systems [J].
Jain, A ;
Nandakumar, K ;
Ross, A .
PATTERN RECOGNITION, 2005, 38 (12) :2270-2285