PHYLOGENETIC REPLAY LEARNING IN DEEP NEURAL NETWORKS

被引:1
|
作者
Glafkides, Jean-Patrice [1 ]
Sher, Gene, I [2 ]
Akdag, Herman [1 ]
机构
[1] PARIS VIII Univ, PARAGRAPHE EA 349, Paris, France
[2] DataValoris, Laramie, WY USA
来源
JORDANIAN JOURNAL OF COMPUTERS AND INFORMATION TECHNOLOGY | 2022年 / 8卷 / 03期
关键词
Neural networks; Neuroevolution; Phylogenetic replay learning; Deep learning; Vanishing gradient;
D O I
10.5455/jjcit.71-1643583878
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Though substantial advancements have been made in training deep neural networks, one problem remains, the vanishing gradient. The very strength of deep neural networks, their depth, is also unfortunately their problem, due to the difficulty of thoroughly training the deeper layers due to the vanishing gradient. This paper proposes "Phylogenetic Replay Learning", a learning methodology that substantially alleviates the vanishing-gradient problem. Unlike the residual learning methods, it does not restrict the structure of the model. Instead, it leverages elements from neuroevolution, transfer learning and layer-by-layer training. We demonstrate that this new approach is able to produce a better performing model and by calculating Shannon entropy of weights, we show that the deeper layers are trained much more thoroughly and contain statistically significantly more information than when a model is trained in a traditional brute force manner.
引用
收藏
页码:218 / 231
页数:14
相关论文
共 50 条
  • [31] Learning fashion compatibility across categories with deep multimodal neural networks
    Sun, Guang-Lu
    He, Jun-Yan
    Wu, Xiao
    Zhao, Bo
    Peng, Qiang
    NEUROCOMPUTING, 2020, 395 (237-246) : 237 - 246
  • [32] Deep Learning Neural Networks Optimization using Hardware Cost Penalty
    Doshi, Rohan
    Hung, Kwok-Wai
    Liang, Luhong
    Chiu, King-Hung
    2016 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2016, : 1954 - 1957
  • [33] NeuroMask: Explaining Predictions of Deep Neural Networks through Mask Learning
    Alzantot, Moustafa
    Widdicombe, Amy
    Julier, Simon
    Srivastava, Mani
    2019 IEEE INTERNATIONAL CONFERENCE ON SMART COMPUTING (SMARTCOMP 2019), 2019, : 81 - 86
  • [34] A Deep Learning Framework for Automated Transfer Learning of Neural Networks
    Balaiah, Thanasekhar
    Jeyadoss, Timothy Jones Thomas
    Thirumurugan, Sainee
    Ravi, Rahul Chander
    2019 11TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTING (ICOAC 2019), 2019, : 428 - 432
  • [35] Adaptive Learning Rate for Unsupervised Learning of Deep Neural Networks
    Golovko, Vladimir
    Mikhno, Egor
    Kroschanka, Aliaksandr
    Chodyka, Marta
    Lichograj, Piotr
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [36] On the Singularity in Deep Neural Networks
    Nitta, Tohru
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT IV, 2016, 9950 : 389 - 396
  • [37] Orthogonal Deep Neural Networks
    Li, Shuai
    Jia, Kui
    Wen, Yuxin
    Liu, Tongliang
    Tao, Dacheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (04) : 1352 - 1368
  • [38] Learning Activation Functions in Deep (Spline) Neural Networks
    Bohra, Pakshal
    Campos, Joaquim
    Gupta, Harshit
    Aziznejad, Shayan
    Unser, Michael
    IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2020, 1 : 295 - 309
  • [39] Stable tensor neural networks for efficient deep learning
    Newman, Elizabeth
    Horesh, Lior
    Avron, Haim
    Kilmer, Misha E.
    FRONTIERS IN BIG DATA, 2024, 7
  • [40] Learning Graph Dynamics using Deep Neural Networks
    Narayan, Apurva
    Roe, Peter H. O'N
    IFAC PAPERSONLINE, 2018, 51 (02): : 433 - 438