A Gradient-Guided Evolutionary Neural Architecture Search

被引:6
|
作者
Xue, Yu [1 ]
Han, Xiaolong [1 ]
Neri, Ferrante [2 ]
Qin, Jiafeng [1 ]
Pelusi, Danilo [3 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Software, Nanjing 210044, Peoples R China
[2] Univ Surrey, Dept Comp Sci, Nat Inspired Comp & Engn Res Grp, Guildford GU2 7XH, England
[3] Univ Teramo, Fac Commun Sci, I-64100 Teramo, Italy
基金
中国国家自然科学基金;
关键词
Computer architecture; Microprocessors; Search problems; Couplings; Evolutionary computation; Encoding; Statistics; gradient optimization; image classification; neural architecture search (NAS);
D O I
10.1109/TNNLS.2024.3371432
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural architecture search (NAS) is a popular method that can automatically design deep neural network structures. However, designing a neural network using NAS is computationally expensive. This article proposes a gradient-guided evolutionary NAS (GENAS) to design convolutional neural networks (CNNs) for image classification. GENAS is a hybrid algorithm that combines evolutionary global and local search operators to evolve a population of subnets sampled from a supernet. Each candidate architecture is encoded as a table describing which operations are associated with the edges between nodes signifying feature maps. Besides, evolutionary optimization uses novel crossover and mutation operators to manipulate the subnets using the proposed tabular encoding. Every n generations, the candidate architectures undergo a local search inspired by differentiable NAS. GENAS is designed to overcome the limitations of both evolutionary and gradient descent NAS. This algorithmic structure enables the performance assessment of the candidate architecture without retraining, thus limiting the NAS calculation time. Furthermore, subnet individuals are decoupled during evaluation to prevent strong coupling of operations in the supernet. The experimental results indicate that the searched structures achieve test errors of 2.45%, 16.86%, and 23.9% on CIFAR-10/100/ImageNet datasets and it costs only 0.26 GPU days on a graphic card. GENAS can effectively expedite the training and evaluation processes and obtain high-performance network structures.
引用
收藏
页码:1 / 13
页数:13
相关论文
共 50 条
  • [31] Neural Architecture Search for Transformers: A Survey
    Chitty-Venkata, Krishna Teja
    Emani, Murali
    Vishwanath, Venkatram
    Somani, Arun K.
    IEEE ACCESS, 2022, 10 : 108374 - 108412
  • [32] Split-Level Evolutionary Neural Architecture Search With Elite Weight Inheritance
    Huang, Junhao
    Xue, Bing
    Sun, Yanan
    Zhang, Mengjie
    Yen, Gary G.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (10) : 13523 - 13537
  • [33] A Novel Training Protocol for Performance Predictors of Evolutionary Neural Architecture Search Algorithms
    Sun, Yanan
    Sun, Xian
    Fang, Yuhan
    Yen, Gary G.
    Liu, Yuqiao
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2021, 25 (03) : 524 - 536
  • [34] A new genetic algorithm based evolutionary neural architecture search for image classification
    Wen, Long
    Gao, Liang
    Li, Xinyu
    Li, Hui
    SWARM AND EVOLUTIONARY COMPUTATION, 2022, 75
  • [35] Component importance preference-based evolutionary graph neural architecture search
    Liu, Yang
    Liu, Jing
    Teng, Yingzhi
    INFORMATION SCIENCES, 2024, 679
  • [36] A new genetic algorithm based evolutionary neural architecture search for image classification
    Wen, Long
    Gao, Liang
    Li, Xinyu
    Li, Hui
    SWARM AND EVOLUTIONARY COMPUTATION, 2022, 75
  • [37] An evolutionary neural architecture search method based on performance prediction and weight inheritance
    Yuan G.
    Xue B.
    Zhang M.
    Information Sciences, 2024, 667
  • [38] TGGS network: A multi-task learning network for gradient-guided knowledge sharing
    Huang, Yongjie
    Han, Xiao
    Chen, Man
    Pan, Zhisong
    KNOWLEDGE-BASED SYSTEMS, 2024, 301
  • [39] GPT-NAS: Neural Architecture Search Meets Generative Pre-Trained Transformer Model
    Yu, Caiyang
    Liu, Xianggen
    Wang, Yifan
    Liu, Yun
    Feng, Wentao
    Deng, Xiong
    Tang, Chenwei
    Lv, Jiancheng
    BIG DATA MINING AND ANALYTICS, 2025, 8 (01): : 45 - 64
  • [40] PASTA: Neural Architecture Search for Anomaly Detection in Multivariate Time Series
    Trirat, Patara
    Lee, Jae-Gil
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024,