Evolutionary Neural Architecture Search with Predictor of Ranking-Based Score

被引:0
作者
Jiang, Peng-Cheng [1 ]
Xue, Yu [1 ]
机构
[1] School of Software, Nanjing University of Information Science and Technology, Nanjing
来源
Jisuanji Xuebao/Chinese Journal of Computers | 2024年 / 47卷 / 11期
关键词
evolutionary computation; genetic algorithms; neural architecture search; ranking prediction; score prediction; surrogate models;
D O I
10.11897/SP.J1016.2024.02522
中图分类号
学科分类号
摘要
The exceptional performance of neural networks has been extensively validated across various practical applications, with architecture serving as the primary determinant of their efficacy. Presently, the state-of-the-art architectures necessitate manual design, heavily relying on expert experience and iterative trial-and-error methodologies for performance validation. In recent years, the emergence of Evolutionary Neural Architecture Search (ENAS) has alleviated the burden associated with manual design. However, despite the ability of ENAS methods to autonomously identify superior architectures, their widespread application remains impeded by the substantial time and computational resources required. Surrogate models can mitigate this excessive resource consumption to some extent. However, existing surrogate model-assisted evolutionary neural architecture searches fail to fully integrate the search and surrogate processes. Moreover, it is difficult for the current surrogate methods to accurately predict network architectural rankings with similar accuracies. Furthermore, existing surrogate models typically necessitate substantial amounts of architectural information as training data to attain satisfactory surrogate accuracy. These limitations hinder the effective assistance of surrogate models in ENAS. thereby constraining its advancement. In this paper, we propose a Rank Score Predictor-assisted Evolutionary Neural Architecture Search method (RSP-ENAS). By introducing a novel loss function specifically designed for rank score prediction, the Multi-Layer Perceptron (MLP) employed as a score predictor can optimally align the ranking of individual performance scores within the population with their actual performance order. During the search process utilizing this method, the predicted scores are directly applicable for elite selection. We introduce a two-stage search strategy in the search phase, initially focusing on accumulating historical information for the surrogate dataset from evaluating a small population and subsequently emphasizing the use of the surrogate model to predict fitness values for a larger population in the later stages. The experiments conducted in this study were performed on the EvoXBench platform, yielding superior results across all benchmark datasets. Additionally, we validated our approach on the ImageNet dataset. Compared to alternative methodologies, our approach successfully identifies the optimal architecture within the NASBench-101 space. On the three datasets within the NASBench-201 space, accuracy improvements of 0. 35%. 1. 12%, and 0. 55% were achieved relative to other optimal methods. In experiments utilizing real datasets on ImageNet, our method demonstrated a 2. 2% enhancement in classification accuracy. Moreover, with the same quantity of data, the ranking results generated by the proposed rank score prediction model exhibited a 1. 55% improvement in Kendall's Tau coefficient when compared to other optimal approaches. We further validated the effectiveness of One-hot encoding and the proposed rank loss within the surrogate model, demonstrating the efficacy of these two components for the overall algorithm. This research underscores the potential of advanced surrogate models to enhance the efficiency and accuracy of neural architecture search processes. By reducing computational costs and improving the precision of architecture rankings, our RSP-ENAS method could significantly advance the practical application and accessibility of neural network design, potentially catalyzing more rapid advancements in the fields of machine learning and artificial intelligence. Future work may explore the surrogate models for less training data, which could yield even more substantial enhancements in neural architecture search methodologies. © 2024 Science Press. All rights reserved.
引用
收藏
页码:2522 / 2535
页数:13
相关论文
共 74 条
[1]  
LeCun Y, Bengio Y, Hinton G., Deep learning, Nature, 521, 7553, pp. 436-444, (2015)
[2]  
Jiao Li-Cheng, Yang Shu-Yuan, Liu Lang, Et al., Seventy-years beyond neural networks: Retrospect and prospect, Chinese Journal of Computers, 39, 8, pp. 1697-1716, (2016)
[3]  
Zhou Fei-Yan, Jin Lin-Peng, Dong Jun, Review of convolu-tional neural network, Chinese Journal of Computers, 40, 6, pp. 1229-1251, (2017)
[4]  
Dou Hui, Zhang Ling-Ming, Han Feng, Et al., Survey on convolutional neural network interpretability, Journal of Software, 35, 1, pp. 159-184, (2024)
[5]  
Li Xiao-Xu, Liu Zhong-Yuan, Wuji -Jie, Et al., Total relation network with attention for few-shot image classification, Chinese Journal of Computers, 46, 2, pp. 371-384, (2023)
[6]  
Yang Xue, Yan Jun-Chi, Visual oriented object detection via feature alignment and Gaussian parameterization, SCIENTIA SINICA Informatioms, 53, 11, pp. 2250-2265, (2023)
[7]  
Tian Xuan, Wang Liang, Ding Qi, Review of image semantic segmentation hased on deep learning, Journal of Software, 30, 2, pp. 440-468, (2019)
[8]  
Liu Y, Sun Y, Xue B, Et al., A survey on evolutionary neural architecture search, IEEE Transactions on Neural Networks and Learning Systems, 34, 2, pp. 550-570, (2023)
[9]  
Meng Zi-Yao, Gu Xue, Liang Yan-Chun, Et al., Deep neural architecture search: A survey, Journal of Computer Research and Development, 58, 1, pp. 22-33, (2021)
[10]  
Zhan Z-H, Shi L, Tan K C, Zhang J., A survey on evolutionary computation for complex continuous optimization, Artificial Intelligence Review, 55, 1, pp. 59-110, (2022)