Evolving Better Initializations For Neural Networks With Pruning

被引:1
作者
Zhou, Ryan [1 ]
Hu, Ting [1 ]
机构
[1] Queens Univ, Kingston, ON, Canada
来源
PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2023 COMPANION | 2023年
关键词
neuroevolution; neural network pruning; neural network initialization; lottery ticket hypothesis; evolution strategies;
D O I
10.1145/3583133.3590745
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent work in deep learning has shown that neural networks can be pruned before training to achieve similar or even better results than training the full network. However, existing pruning methods are limited and do not necessarily yield optimal solutions. In this work, we show that perturbing the network by re-initializing the pruned weights and re-pruning can improve performance. We propose to iteratively re-initialize and re-prune using a hill climbing (1 + 1) evolution strategy. We examine the cause of these improvements and show that this method can consistently improve the subnetwork without increasing its size, pointing to a potential new application of evolutionary computing in deep learning.
引用
收藏
页码:703 / 706
页数:4
相关论文
共 7 条
[1]   Reconciling modern machine-learning practice and the classical bias-variance trade-off [J].
Belkin, Mikhail ;
Hsu, Daniel ;
Ma, Siyuan ;
Mandal, Soumik .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2019, 116 (32) :15849-15854
[2]  
Fischer J, 2022, 10 INT C LEARNING RE
[3]  
Frankle J., 2021, INT C LEARN REPR
[4]  
Frankle Jonathan, 2019, PHYSIOTHER THEOR PR
[5]   What's Hidden in a Randomly Weighted Neural Network? [J].
Ramanujan, Vivek ;
Wortsman, Mitchell ;
Kembhavi, Aniruddha ;
Farhadi, Ali ;
Rastegari, Mohammad .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :11890-11899
[6]  
Wang H, 2022, PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, P5638
[7]  
Zhou Hattie, 2019, Advances in Neural Information Processing Systems,, V32