SWANN: Small-World Architecture for Fast Convergence of Neural Networks

被引:3
|
作者
Javaheripi, Mojan [1 ]
Rouhani, Bita Darvish [2 ]
Koushanfar, Farinaz [1 ]
机构
[1] Univ Calif San Diego, Dept Elect & Comp Engn, San Diego, CA 92093 USA
[2] Microsoft, Redmond, WA 98052 USA
关键词
Deep learning; on-device training; small-world networks; PERFORMANCE; CONSENSUS;
D O I
10.1109/JETCAS.2021.3125309
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
On-device intelligence has become increasingly widespread in the modern smart application landscape. A standing challenge for the applicability of on-device intelligence is the excessively high computation cost of training highly accurate Deep Learning (DL) models. These models require a large number of training iterations to reach a high convergence accuracy, hindering their applicability to resource-constrained embedded devices. This paper proposes a novel transformation which changes the topology of the DL architecture to reach an optimal cross-layer connectivity. This, in turn, significantly reduces the number of training iterations required for reaching a target accuracy. Our transformation leverages the important observation that for a set level of accuracy, convergence is fastest when network topology reaches the boundary of a Small-World Network. Small-world graphs are known to possess a specific connectivity structure that enables enhanced signal propagation among nodes. Our small-world models, called SWANNs, provide several intriguing benefits: they facilitate data (gradient) flow within the network, enable feature-map reuse by adding long-range connections and accommodate various network architectures/datasets. Compared to densely connected networks (e.g., DenseNets), SWANNs require a substantially fewer number of training parameters while maintaining a similar level of classification accuracy. We evaluate our networks on various DL model architectures and image classification datasets, namely, MNIST, CIFAR10, CIFAR100, and ImageNet. Our experiments demonstrate an average of approximate to 2.1 x improvement in convergence speed to the desired accuracy.
引用
收藏
页码:575 / 585
页数:11
相关论文
共 50 条
  • [11] Effect of intermodular connection on fast sparse synchronization in clustered small-world neural networks
    Kim, Sang-Yoon
    Lim, Woochang
    PHYSICAL REVIEW E, 2015, 92 (05):
  • [12] Cellular neural networks can mimic small-world networks
    Yang, T
    Chua, LO
    INTERNATIONAL JOURNAL OF BIFURCATION AND CHAOS, 1999, 9 (10): : 2105 - 2126
  • [13] Fault Tolerant Small-World Cellular Neural Networks
    Matsumoto, Katsuyoshi
    Nomoto, Shingo
    Uehara, Minoru
    Mori, Hideki
    2008 22ND INTERNATIONAL WORKSHOPS ON ADVANCED INFORMATION NETWORKING AND APPLICATIONS, VOLS 1-3, 2008, : 168 - 172
  • [14] Coupling parameter in synchronization of small-world neural networks
    Wang, SJ
    Xu, XJ
    Wu, ZX
    Wang, YH
    INTERNATIONAL JOURNAL OF MODERN PHYSICS C, 2005, 16 (12): : 1841 - 1848
  • [15] Stochastic resonance in coupled small-world neural networks
    Yu Hai-Tao
    Wang Jiang
    Liu Chen
    Che Yan-Qiu
    Deng Bin
    Wei Xi-Le
    ACTA PHYSICA SINICA, 2012, 61 (06)
  • [16] Coherence resonance in neural networks with small-world connections
    Zhou Xiao-Rong
    Luo Xiao-Shu
    ACTA PHYSICA SINICA, 2008, 57 (05) : 2849 - 2853
  • [17] Fault Tolerance for Small-World Cellular Neural Networks
    Matsumoto, Kautsuyoshi
    Uehara, Minoru
    Mori, Hideki
    NETWORK-BASED INFORMATION SYSTEMS, PROCEEDINGS, 2008, 5186 : 223 - 231
  • [18] Small-World Outer Synchronization of Small-World Chaotic Networks
    Arellano-Delgado, A.
    Lopez-Gutierrez, R. M.
    Martinez-Clark, R.
    Cruz-Hernandez, C.
    JOURNAL OF COMPUTATIONAL AND NONLINEAR DYNAMICS, 2018, 13 (10):
  • [19] Fast response and temporal coherent oscillations in small-world networks
    Lago-Fernández, LF
    Huerta, R
    Corbacho, F
    Sigüenza, JA
    PHYSICAL REVIEW LETTERS, 2000, 84 (12) : 2758 - 2761
  • [20] The influence of autapse on synchronous firing in small-world neural networks
    Peng, Lu
    Tang, Jun
    Ma, Jun
    Luo, Jinming
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2022, 594