Subaging in underparametrized deep neural networks

被引:1
作者
Herrera Segura, Carolina [1 ]
Montoya, Edison [2 ,4 ]
Tapias, Diego [3 ]
机构
[1] Univ Antioquia, Inst Fis, Medellin, Colombia
[2] BCFort, Medellin, Colombia
[3] Univ Gottingen, Inst Theoret Phys, Gottingen, Germany
[4] Univ Antioquia, Medellin, Colombia
来源
MACHINE LEARNING-SCIENCE AND TECHNOLOGY | 2022年 / 3卷 / 03期
关键词
subaging; deep neural networks; glassy dynamics; underparametrized; STATISTICAL-MECHANICS; REGIMES; ENERGY;
D O I
10.1088/2632-2153/ac8f1b
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider a simple classification problem to show that the dynamics of finite-width Deep Neural Networks in the underparametrized regime gives rise to effects similar to those associated with glassy systems, namely a slow evolution of the loss function and aging. Remarkably, the aging is sublinear in the waiting time (subaging) and the power-law exponent characterizing it is robust to different architectures under the constraint of a constant total number of parameters. Our results are maintained in the more complex scenario of the MNIST database. We find that for this database there is a unique exponent ruling the subaging behavior in the whole phase.
引用
收藏
页数:10
相关论文
共 50 条
[41]   Federated Repair of Deep Neural Networks [J].
Li Calsi, Davide ;
Laurent, Thomas ;
Arcaini, Paolo ;
Ishikawa, Fuyuki .
PROCEEDINGS OF THE 2024 IEEE/ACM INTERNATIONAL WORKSHOP ON DEEP LEARNING FOR TESTING AND TESTING FOR DEEP LEARNING, DEEPTEST 2024, 2024, :17-24
[42]   THE COMBINATION OF CONVOLUTION NEURAL NETWORKS AND DEEP NEURAL NETWORKS FOR FAKE NEWS DETECTION [J].
Jawad, Zainab A. ;
Obaid, Ahmed J. .
JOURNAL OF ENGINEERING SCIENCE AND TECHNOLOGY, 2023, 18 (01) :814-826
[43]   Attacking Neural Networks with Neural Networks: Towards Deep Synchronization for Backdoor Attacks [J].
Guan, Zihan ;
Sun, Lichao ;
Du, Mengnan ;
Liu, Ninghao .
PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, :608-618
[44]   Deep Learning with Darwin: Evolutionary Synthesis of Deep Neural Networks [J].
Mohammad Javad Shafiee ;
Akshaya Mishra ;
Alexander Wong .
Neural Processing Letters, 2018, 48 :603-613
[45]   Deep Energy: Task Driven Training of Deep Neural Networks [J].
Golts, Alona ;
Freedman, Daniel ;
Elad, Michael .
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2021, 15 (02) :324-338
[46]   Deep Learning with Darwin: Evolutionary Synthesis of Deep Neural Networks [J].
Shafiee, Mohammad Javad ;
Mishra, Akshaya ;
Wong, Alexander .
NEURAL PROCESSING LETTERS, 2018, 48 (01) :603-613
[47]   Deep Kronecker neural networks: A general framework for neural networks with adaptive activation functions [J].
Jagtap, Ameya D. ;
Shin, Yeonjong ;
Kawaguchi, Kenji ;
Karniadakis, George Em .
NEUROCOMPUTING, 2022, 468 (165-180) :165-180
[48]   Lessons From Deep Neural Networks for Studying the Coding Principles of Biological Neural Networks [J].
Bae, Hyojin ;
Kim, Sang Jeong ;
Kim, Chang-Eop .
FRONTIERS IN SYSTEMS NEUROSCIENCE, 2021, 14
[49]   Deep Neural Networks for Cooperative Lidar Localization in Vehicular Networks [J].
Barbieri, Luca ;
Brambilla, Mattia ;
Nicoli, Monica .
ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, :185-190
[50]   A Survey of Attacks and Defenses for Deep Neural Networks [J].
Machooka, Daniel ;
Yuan, Xiaohong ;
Esterline, Albert .
2023 IEEE INTERNATIONAL CONFERENCE ON CYBER SECURITY AND RESILIENCE, CSR, 2023, :254-261