Data-specific activation function learning for constructive neural networks

被引:0
|
作者
Xia, Zhenxing [1 ,2 ]
Dai, Wei [1 ,2 ]
Liu, Xin [2 ]
Zhang, Haijun [3 ]
Ma, Xiaoping [1 ,2 ]
机构
[1] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221116, Peoples R China
[2] China Univ Min & Technol, Artificial Intelligence Res Inst, Xuzhou 221116, Peoples R China
[3] Harbin Inst Technol, Shenzhen Grad Sch, Dept Comp Sci, Shenzhen 518055, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Activation function; Self-learning strategy; Reward-penalty mechanism; Constructive neural networks; Data-specific; ACCURACY;
D O I
10.1016/j.neucom.2024.129020
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Activation functions play a crucial role in learning and expressive capabilities of advanced neural networks due to their non-linear or non-saturated properties. However, how to determine the appropriate activation function from various candidates is a challenging yet not well-addressed topic. To address the issue, a novel self-learning approach, called as data-specific activation function learning (DSAFL) algorithm, is proposed to establish constructive neural network on one-time by adaptively selecting appropriate activation function based on the specific data characteristics. To assess the space dimension mapping abilities of different activation functions, the configuration probabilities are used to guide the generation of various candidate activation functions and corresponding candidate hidden node. In the learning stage, an exploration-exploitation mechanism composed of the random algorithm and the greedy strategy is developed to obtain the influence of different candidate activation functions, thereby avoiding configuration probabilities falling into local optimum. A reward-penalty mechanism is built to update the configuration probabilities and enhance the robustness of network by integrating the simulated annealing strategy. In final, the activation function with the highest configuration probability, as the best one, is used to reconstruct the neural network. Experimental results on both regression and classification tasks demonstrate the efficiency and effectiveness of DSAFL in the activation function selection problems of a class of constructive neural networks.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Constructive function approximation by neural networks with optimized activation functions and fixed weights
    Li, Feng-Jun
    NEURAL COMPUTING & APPLICATIONS, 2019, 31 (09): : 4613 - 4628
  • [2] Constructive function approximation by neural networks with optimized activation functions and fixed weights
    Feng-Jun Li
    Neural Computing and Applications, 2019, 31 : 4613 - 4628
  • [3] A Constructive Learning Algorithm Based on Division of Training Data for Multilayer Neural Networks
    Uno, T.
    Koakutsu, S.
    Hirata, H.
    Denki Gakkai Ronbunshi. C, Erekutoronikusu Joho Kogaku, Shisutemu, 116 (10):
  • [5] Continual robot learning with constructive neural networks
    Grossmann, A
    Poli, R
    LEARNING ROBOTS, PROCEEDINGS, 1998, 1545 : 95 - 108
  • [6] Data-Specific Analysis of String Sorting
    Seidel, Raimund
    PROCEEDINGS OF THE TWENTY-FIRST ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, 2010, 135 : 1278 - 1286
  • [7] Constructive approximation to real function by wavelet neural networks
    Hou Muzhou
    Han Xuli
    Gan Yixuan
    Neural Computing and Applications, 2009, 18 : 883 - 889
  • [8] Constructive approximation to real function by wavelet neural networks
    Hou Muzhou
    Han Xuli
    Gan Yixuan
    NEURAL COMPUTING & APPLICATIONS, 2009, 18 (08): : 883 - 889
  • [9] Incremental Feature Learning Using Constructive Neural Networks
    Sadreddin, Armin
    Sadaoui, Samira
    2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 704 - 708
  • [10] Data-specific Feature Point Descriptor Matching Using Dictionary Learning and Graphical Models
    Guerrero, Ricardo
    Rueckert, Daniel
    MEDICAL IMAGING 2013: IMAGE PROCESSING, 2013, 8669