SaDENAS: A self-adaptive differential evolution algorithm for neural architecture search

被引:5
作者
Han, Xiaolong [1 ]
Xue, Yu [1 ]
Wang, Zehong [2 ]
Zhang, Yong [3 ]
Muravev, Anton [4 ]
Gabbouj, Moncef [4 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Software, Nanjing, Peoples R China
[2] Univ Notre Dame, Dept Comp Sci & Engn, Notre Dame, IN USA
[3] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou, Peoples R China
[4] Tampere Univ, Fac Informat Technol & Commun Sci, Tampere, Finland
关键词
Neural architecture search; Evolutionary computation; Differential evolution; Differentiable architecture search;
D O I
10.1016/j.swevo.2024.101736
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Evolutionary neural architecture search (ENAS) and differentiable architecture search (DARTS) are all prominent algorithms in neural architecture search, enabling the automated design of deep neural networks. To leverage the strengths of both methods, there exists a framework called continuous ENAS, which alternates between using gradient descent to optimize the supernet and employing evolutionary algorithms to optimize the architectural encodings. However, in continuous ENAS, there exists a premature convergence issue accompanied by the small model trap, which is a common issue in NAS. To address this issue, this paper proposes a self-adaptive differential evolution algorithm for neural architecture search (SaDENAS), which can reduce the interference caused by small models to other individuals during the optimization process, thereby avoiding premature convergence. Specifically, SaDENAS treats architectures within the search space as architectural encodings, leveraging vector differences between encodings as the basis for evolutionary operators. To achieve a trade-off between exploration and exploitation, we integrate both local and global search strategies with a mutation scaling factor to adaptively balance these two strategies. Empirical findings demonstrate that our proposed algorithm achieves better performance with superior convergence compared to other algorithms.
引用
收藏
页数:10
相关论文
共 54 条
[1]  
Baker B., 2017, ICLR
[2]  
Cai ZC, 2024, AAAI CONF ARTIF INTE, P11159
[3]  
Chen XN, 2020, PR MACH LEARN RES, V119
[4]   Progressive Differentiable Architecture Search: Bridging the Depth Gap between Search and Evaluation [J].
Chen, Xin ;
Xie, Lingxi ;
Wu, Jun ;
Tian, Qi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1294-1303
[5]  
Chu X., 2021, INT C LEARN REPR
[6]  
Chu X., 2020, EUR C COMP VIS
[7]   A Cell-Based Fast Memetic Algorithm for Automated Convolutional Neural Architecture Design [J].
Dong, Junwei ;
Hou, Boyu ;
Feng, Liang ;
Tang, Huajin ;
Tan, Kay Chen ;
Ong, Yew-Soon .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) :9040-9053
[8]   EMQ: Evolving Training-free Proxies for Automated Mixed Precision Quantization [J].
Dong, Peijie ;
Li, Lujun ;
Wei, Zimian ;
Niu, Xin ;
Tian, Zhiliang ;
Pan, Hengyue .
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, :17030-17040
[9]   Random and Coherent Noise Suppression in DAS-VSP Data by Using a Supervised Deep Learning Method [J].
Dong, Xintong ;
Li, Yue ;
Zhong, Tie ;
Wu, Ning ;
Wang, Hongzhou .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
[10]   One-Shot Neural Architecture Search via Self-Evaluated Template Network [J].
Dong, Xuanyi ;
Yang, Yi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3680-3689