Hierarchical multi-scale parametric optimization of deep neural networks

被引:2
作者
Zhang, Sushen [1 ]
Vassiliadis, Vassilios S. [2 ]
Dorneanu, Bogdan [3 ]
Arellano-Garcia, Harvey [3 ]
机构
[1] Univ Cambridge, Dept Chem Engn & Biotechnol, Cambridge, England
[2] Cambridge Simulat Solut LTD, Cambridge, England
[3] Brandenburg Univ Technol Cottbus Senftenberg, LS Prozess & Anlagentech, Cottbus, Germany
关键词
Deep neural networks; Hierarchical multi-scale search; Scaling factor; Sensitivity analysis; Finite difference; Automatic differentiation; SENSITIVITY-ANALYSIS; PREDICTION; BACKPROPAGATION;
D O I
10.1007/s10489-023-04745-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traditionally, sensitivity analysis has been utilized to determine the importance of input variables to a deep neural network (DNN). However, the quantification of sensitivity for each neuron in a network presents a significant challenge. In this article, a selective method for calculating neuron sensitivity in layers of neurons concerning network output is proposed. This approach incorporates scaling factors that facilitate the evaluation and comparison of neuron importance. Additionally, a hierarchical multi-scale optimization framework is proposed, where layers with high-importance neurons are selectively optimized. Unlike the traditional backpropagation method that optimizes the whole network at once, this alternative approach focuses on optimizing the more important layers. This paper provides fundamental theoretical analysis and motivating case study results for the proposed neural network treatment. The framework is shown to be effective in network optimization when applied to simulated and UCI Machine Learning Repository datasets. This alternative training generates local minima close to or even better than those obtained with the backpropagation method, utilizing the same starting points for comparative purposes within a multi-start optimization procedure. Moreover, the proposed approach is observed to be more efficient for large-scale DNNs. These results validate the proposed algorithmic framework as a rigorous and robust new optimization methodology for training (fitting) neural networks to input/output data series of any given system.
引用
收藏
页码:24963 / 24990
页数:28
相关论文
共 85 条
  • [1] Advanced metaheuristic optimization techniques in applications of deep neural networks: a review
    Abd Elaziz, Mohamed
    Dahou, Abdelghani
    Abualigah, Laith
    Yu, Liyang
    Alshinwan, Mohammad
    Khasawneh, Ahmad M.
    Lu, Songfeng
    [J]. NEURAL COMPUTING & APPLICATIONS, 2021, 33 (21) : 14079 - 14099
  • [2] Artificial Neural Networks Based Optimization Techniques: A Review
    Abdolrasol, Maher G. M.
    Hussain, S. M. Suhail
    Ustun, Taha Selim
    Sarker, Mahidur R.
    Hannan, Mahammad A.
    Mohamed, Ramizi
    Ali, Jamal Abd
    Mekhilef, Saad
    Milad, Abdalrhman
    [J]. ELECTRONICS, 2021, 10 (21)
  • [3] Tuning Deep Neural Networks for Predicting Energy Consumption in Arid Climate Based on Buildings Characteristics
    Al-Shargabi, Amal A.
    Almhafdy, Abdulbasit
    Ibrahim, Dina M.
    Alghieth, Manal
    Chiclana, Francisco
    [J]. SUSTAINABILITY, 2021, 13 (22)
  • [4] Application of a Machine Learning Algorithm for Evaluation of Stiff Fractional Modeling of Polytropic Gas Spheres and Electric Circuits
    Alarfaj, Fawaz Khaled
    Khan, Naveed Ahmad
    Sulaiman, Muhammad
    Alomair, Abdullah M.
    [J]. SYMMETRY-BASEL, 2022, 14 (12):
  • [5] Review of deep learning: concepts, CNN architectures, challenges, applications, future directions
    Alzubaidi, Laith
    Zhang, Jinglan
    Humaidi, Amjad J.
    Al-Dujaili, Ayad
    Duan, Ye
    Al-Shamma, Omran
    Santamaria, J.
    Fadhel, Mohammed A.
    Al-Amidie, Muthana
    Farhan, Laith
    [J]. JOURNAL OF BIG DATA, 2021, 8 (01)
  • [6] Sensitivity analysis for interpretation of machine learning based segmentation models in cardiac MRI
    Ankenbrand, Markus J.
    Shainberg, Liliia
    Hock, Michael
    Lohr, David
    Schreiber, Laura M.
    [J]. BMC MEDICAL IMAGING, 2021, 21 (01)
  • [7] Hardware Approximate Techniques for Deep Neural Network Accelerators: A Survey
    Armeniakos, Giorgos
    Zervakis, Georgios
    Soudris, Dimitrios
    Henkel, Joerg
    [J]. ACM COMPUTING SURVEYS, 2023, 55 (04)
  • [8] Aszemi NM, 2019, INT J ADV COMPUT SC, V10, P269
  • [9] On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation
    Bach, Sebastian
    Binder, Alexander
    Montavon, Gregoire
    Klauschen, Frederick
    Mueller, Klaus-Robert
    Samek, Wojciech
    [J]. PLOS ONE, 2015, 10 (07):
  • [10] Gradient Amplification: An Efficient Way to Train Deep Neural Networks
    Basodi, Sunitha
    Ji, Chunyan
    Zhang, Haiping
    Pan, Yi
    [J]. BIG DATA MINING AND ANALYTICS, 2020, 3 (03) : 196 - 207