RelativeNAS: Relative Neural Architecture Search via Slow-Fast Learning

被引:29
|
作者
Tan, Hao [1 ]
Cheng, Ran [1 ]
Huang, Shihua [1 ]
He, Cheng [1 ]
Qiu, Changxiao [2 ]
Yang, Fan [2 ]
Luo, Ping [3 ]
机构
[1] Southern Univ Sci & Technol, Univ Key Lab Evolving Intelligent Syst Guangdong, Dept Comp Sci & Engn, Shenzhen 518055, Peoples R China
[2] Huawei Technol Co Ltd, Hisilicon Res Dept, Shenzhen 518055, Peoples R China
[3] Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Computer architecture; Statistics; Sociology; Search problems; Optimization; Neural networks; Estimation; AutoML; convolutional neural network (CNN); neural architecture search (NAS); population-based search; slow-fast learning; NETWORKS;
D O I
10.1109/TNNLS.2021.3096658
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the remarkable successes of convolutional neural networks (CNNs) in computer vision, it is time-consuming and error-prone to manually design a CNN. Among various neural architecture search (NAS) methods that are motivated to automate designs of high-performance CNNs, the differentiable NAS and population-based NAS are attracting increasing interests due to their unique characters. To benefit from the merits while overcoming the deficiencies of both, this work proposes a novel NAS method, RelativeNAS. As the key to efficient search, RelativeNAS performs joint learning between fast learners (i.e., decoded networks with relatively lower loss value) and slow learners in a pairwise manner. Moreover, since RelativeNAS only requires low-fidelity performance estimation to distinguish each pair of fast learner and slow learner, it saves certain computation costs for training the candidate architectures. The proposed RelativeNAS brings several unique advantages: 1) it achieves state-of-the-art performances on ImageNet with top-1 error rate of 24.88%, that is, outperforming DARTS and AmoebaNet-B by 1.82% and 1.12%, respectively; 2) it spends only 9 h with a single 1080Ti GPU to obtain the discovered cells, that is, 3.75x and 7875x faster than DARTS and AmoebaNet, respectively; and 3) it provides that the discovered cells obtained on CIFAR-10 can be directly transferred to object detection, semantic segmentation, and keypoint detection, yielding competitive results of 73.1% mAP on PASCAL VOC, 78.7% mIoU on Cityscapes, and 68.5% AP on MSCOCO, respectively. The implementation of RelativeNAS is available at https://github.com/EMI-Group/RelativeNAS.
引用
收藏
页码:475 / 489
页数:15
相关论文
共 50 条
  • [31] FINCH: Enhancing Federated Learning With Hierarchical Neural Architecture Search
    Liu, Jianchun
    Yan, Jiaming
    Xu, Hongli
    Wang, Zhiyuan
    Huang, Jinyang
    Xu, Yang
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (05) : 6012 - 6026
  • [32] Multiobjective Reinforcement Learning-Based Neural Architecture Search for Efficient Portrait Parsing
    Lyu, Bo
    Wen, Shiping
    Shi, Kaibo
    Huang, Tingwen
    IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (02) : 1158 - 1169
  • [33] Efficient Self-learning Evolutionary Neural Architecture Search
    Qiu, Zhengzhong
    Bi, Wei
    Xu, Dong
    Guo, Hua
    Ge, Hongwei
    Liang, Yanchun
    Lee, Heow Pueh
    Wu, Chunguo
    APPLIED SOFT COMPUTING, 2023, 146
  • [34] Metaheuristics and machine learning: an approach with reinforcement learning assisting neural architecture search
    Venske, Sandra Mara Scos
    de Almeida, Carolina Paula
    Delgado, Myriam Regattieri
    JOURNAL OF HEURISTICS, 2024, 30 (3-4) : 199 - 224
  • [35] Neural architecture search via progressive partial connection with attention mechanism
    Jin, Cong
    Huang, Jinjie
    Chen, Yuanjian
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [36] Peaches: Personalized Federated Learning With Neural Architecture Search in Edge Computing
    Yan, Jiaming
    Liu, Jianchun
    Xu, Hongli
    Wang, Zhiyuan
    Qiao, Chunming
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (11) : 10296 - 10312
  • [37] Building High-Throughput Neural Architecture Search Workflows via a Decoupled Fitness Prediction Engine
    Keller Rorabaugh, Ariel
    Caino-Lores, Silvina
    Johnston, Travis
    Taufer, Michela
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (11) : 2913 - 2926
  • [38] EEG-Based Sleep Stage Classification via Neural Architecture Search
    Kong, Gangwei
    Li, Chang
    Peng, Hu
    Han, Zhihui
    Qiao, Heyuan
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2023, 31 : 1075 - 1085
  • [39] EEG-based Emotion Recognition via Transformer Neural Architecture Search
    Li, Chang
    Zhang, Zhongzhen
    Zhang, Xiaodong
    Huang, Guoning
    Liu, Yu
    Chen, Xun
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (04) : 6016 - 6025
  • [40] Contrastive meta-reinforcement learning for heterogeneous graph neural architecture search
    Xu, Zixuan
    Wu, Jia
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 260