Efficient perturbation-aware distinguishing score for zero-shot neural architecture search

被引:0
作者
Huang, Junhao [1 ]
Xue, Bing [1 ]
Sun, Yanan [2 ]
Zhang, Mengjie [1 ]
Yen, Gary G. [2 ]
机构
[1] Centre for Data Science and Artificial Intelligence & School of Engineering and Computer Science, Victoria University of Wellington, Wellington
[2] School of Computer Science, Sichuan University, Chengdu
关键词
Activation patterns; Deep neural networks; Neural architecture search; Zero-cost proxy;
D O I
10.1016/j.asoc.2025.113447
中图分类号
学科分类号
摘要
Zero-cost proxies are under the spotlight of neural architecture search (NAS) lately, thanks to their low computational cost in predicting architecture performance in a training-free manner. The NASWOT score is one of the representative proxies that measures the architecture's ability to distinguish inputs at the activation layers. However, obtaining such a score still requires considerable calculations on a large kernel matrix about input similarity. Moreover, the NASWOT score is relatively coarse-grained and provides a rough estimation of the architecture's ability to distinguish general inputs. In this paper, to further reduce the computational complexity, we first propose a simplified NASWOT scoring term by relaxing its original matrix-based calculation into a vector-based one. More importantly, we develop a fine-grained perturbation-aware term to measure how well the architecture can distinguish between inputs and their perturbed counterparts. We propose a layer-wise score multiplication approach to combine these two scoring terms, deriving a new proxy, named efficient perturbation-aware distinguishing score (ePADS). Experiments on various NAS spaces and datasets show that ePADS consistently outperforms other zero-cost proxies in terms of both predictive reliability and efficiency. Particularly, ePADS achieves the highest ranking correlation among the advanced competitors (e.g., Kendall's coefficient of 0.620 on NAS-Bench-201 with ImageNet-16-120 and 0.485 on NDS-ENAS), and ePADS-based random architecture search spends only 0.018 GPU days on DARTS-CIFAR to find networks with an average error rate of 2.64%. © 2025 Elsevier B.V.
引用
收藏
相关论文
共 52 条
[1]  
White C., Safari M., Sukthanker R., Ru B., Elsken T., Zela A., Dey D., Hutter F., Neural architecture search: Insights from 1000 papers, (2023)
[2]  
Liu H., Simonyan K., Yang Y., DARTS: Differentiable Architecture Search, pp. 1-13, (2019)
[3]  
Tan M., Chen B., Pang R., Vasudevan V., Sandler M., Howard A., Le Q.V., MnasNet: Platform-Aware Neural Architecture Search for Mobile, pp. 2820-2828, (2019)
[4]  
Cai H., Gan C., Wang T., Zhang Z., Han S., Once-for-All: Train One Network and Specialize it for Efficient Deployment, pp. 1-14, (2020)
[5]  
Huang J., Xue B., Sun Y., Zhang M., Yen G.G., Split-level evolutionary neural architecture search with elite weight inheritance, IEEE Trans. Neural Networks Learn. Syst., 35, 10, pp. 13523-13537, (2024)
[6]  
Real E., Moore S., Selle A., Saxena S., Suematsu Y.L., Tan J., Le Q.V., Kurakin A., Large-scale evolution of image classifiers, International Conference on Machine Learning, ICML, pp. 2902-2911, (2017)
[7]  
Zoph B., Vasudevan V., Shlens J., Le Q.V., Learning Transferable Architectures for Scalable Image Recognition, pp. 8697-8710, (2018)
[8]  
Zhou D., Zhou X., Zhang W., Loy C.C., Yi S., Zhang X., Ouyang W., EcoNAS: Finding proxies for economical neural architecture search, pp. 11396-11404, (2020)
[9]  
Huang J., Xue B., Sun Y., Zhang M., Yen G.G., Particle swarm optimization for compact neural architecture search for image classification, IEEE Trans. Evol. Comput., 27, 5, pp. 1298-1312, (2023)
[10]  
Deng B., Yan J., Lin D., Peephole: Predicting network performance before training, (2017)