SWANN: Small-World Architecture for Fast Convergence of Neural Networks

被引:3
|
作者
Javaheripi, Mojan [1 ]
Rouhani, Bita Darvish [2 ]
Koushanfar, Farinaz [1 ]
机构
[1] Univ Calif San Diego, Dept Elect & Comp Engn, San Diego, CA 92093 USA
[2] Microsoft, Redmond, WA 98052 USA
关键词
Deep learning; on-device training; small-world networks; PERFORMANCE; CONSENSUS;
D O I
10.1109/JETCAS.2021.3125309
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
On-device intelligence has become increasingly widespread in the modern smart application landscape. A standing challenge for the applicability of on-device intelligence is the excessively high computation cost of training highly accurate Deep Learning (DL) models. These models require a large number of training iterations to reach a high convergence accuracy, hindering their applicability to resource-constrained embedded devices. This paper proposes a novel transformation which changes the topology of the DL architecture to reach an optimal cross-layer connectivity. This, in turn, significantly reduces the number of training iterations required for reaching a target accuracy. Our transformation leverages the important observation that for a set level of accuracy, convergence is fastest when network topology reaches the boundary of a Small-World Network. Small-world graphs are known to possess a specific connectivity structure that enables enhanced signal propagation among nodes. Our small-world models, called SWANNs, provide several intriguing benefits: they facilitate data (gradient) flow within the network, enable feature-map reuse by adding long-range connections and accommodate various network architectures/datasets. Compared to densely connected networks (e.g., DenseNets), SWANNs require a substantially fewer number of training parameters while maintaining a similar level of classification accuracy. We evaluate our networks on various DL model architectures and image classification datasets, namely, MNIST, CIFAR10, CIFAR100, and ImageNet. Our experiments demonstrate an average of approximate to 2.1 x improvement in convergence speed to the desired accuracy.
引用
收藏
页码:575 / 585
页数:11
相关论文
共 50 条
  • [1] Systematic rewiring in associative neural networks with small-world architecture
    Dekhtyarenko, OK
    Proceedings of the International Joint Conference on Neural Networks (IJCNN), Vols 1-5, 2005, : 1178 - 1181
  • [2] Emergence of the small-world architecture in neural networks by activity dependent growth
    Gafarov, F. M.
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2016, 461 : 409 - 418
  • [3] Fast and robust image segmentation by small-world neural oscillator networks
    Li, Chunguang
    Li, Yuke
    COGNITIVE NEURODYNAMICS, 2011, 5 (02) : 209 - 220
  • [4] Fast and robust image segmentation by small-world neural oscillator networks
    Chunguang Li
    Yuke Li
    Cognitive Neurodynamics, 2011, 5 : 209 - 220
  • [5] Fast Synchronization with Directed Small-World Networks
    Montalbano, Flavio
    Khodaverdian, Saman
    Adamy, Juergen
    2015 12TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, COMPUTING SCIENCE AND AUTOMATIC CONTROL (CCE 2015), 2015,
  • [6] Fast Similarity Search in Small-World Networks
    Aoyama, Kazuo
    Saito, Kazumi
    Yamada, Takeshi
    Ueda, Naonori
    COMPLEX NETWORKS, 2009, 207 : 185 - +
  • [7] Synchronization in Small-World Chaotic Neural Networks
    Shinkai, Tsuyoshi
    Adachi, Masaharu
    2011 INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS AND NEURAL COMPUTING (FSNC 2011), VOL I, 2011, : 156 - 159
  • [8] Structured information in small-world neural networks
    Dominguez, David
    Gonzalez, Mario
    Serrano, Eduardo
    Rodriguez, Francisco B.
    PHYSICAL REVIEW E, 2009, 79 (02):
  • [9] Fastest learning in small-world neural networks
    Simard, D
    Nadeau, L
    Kröger, H
    PHYSICS LETTERS A, 2005, 336 (01) : 8 - 15
  • [10] Stability of small-world networks of neural populations
    Gray, R. T.
    Fung, C. K. C.
    Robinson, P. A.
    NEUROCOMPUTING, 2009, 72 (7-9) : 1565 - 1574