Finding a succinct multi-layer perceptron having shared weights

被引:0
|
作者
Tanahashi, Y [1 ]
Chin, XF [1 ]
Saito, K [1 ]
Nakano, R [1 ]
机构
[1] Nagoya Inst Technol, Showa Ku, Nagoya, Aichi 4668555, Japan
来源
PROCEEDINGS OF THE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), VOLS 1-5 | 2005年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a method to find a succinct neural network having shared weights. We focus on weight sharing. Weight sharing constrains the freedom of weight values and weights are allowed to have one of common weights. A near-zero common weight can be eliminated, called weight pruning. Recently, a weight sharing method called BCW has been proposed. The BCW employs merge and split operations based on 2nd-order optimal criteria, and can escape local optima through bidirectional clustering. However, the BCW assumes a vital network parameter J, the number of hidden units, is given. This paper modifies the BCW to make the procedure faster so that the selection of J based on cross-validation can be done in reasonable cpu time. Our experiments showed that the proposed method can restore the original model for an artificial data set, and finds a small number of common weights and an interesting tendency for a real data set.
引用
收藏
页码:1418 / 1423
页数:6
相关论文
共 50 条
  • [1] Finding nominally conditioned multivariate polynomials using a four-layer perceptron having shared weights
    Tanahashi, Yusuke
    Saito, Kazumi
    Kitakoshi, Daisuke
    Nakano, Ryohei
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 2, PROCEEDINGS, 2006, 4252 : 969 - 976
  • [2] Efficient Estimation for Shared Latent Space Using Multi-layer Perceptron
    Ohyama, Mariho
    Kobayashi, Ichiro
    2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 2425 - 2430
  • [3] Multi-Layer Perceptron with Pulse Glial Chain Having Oscillatory Excitation Threshold
    Ikuta, Chihiro
    Uwate, Yoko
    Nishio, Yoshifumi
    2015 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2015, : 1330 - 1333
  • [4] Hybrid wolf-bat algorithm for optimization of connection weights in multi-layer perceptron
    Agrawal, Utkarsh
    Arora, Jatin
    Singh, Rahul
    Gupta, Deepak
    Khanna, Ashish
    Khamparia, Aditya
    Gupta, Deepak (deepakgupta@mait.ac.in), 1600, Association for Computing Machinery, 2 Penn Plaza, Suite 701, New York, NY 10121-0701, United States (16):
  • [5] Hybrid Wolf-Bat Algorithm for Optimization of Connection Weights in Multi-layer Perceptron
    Agrawal, Utkarsh
    Arora, Jatin
    Singh, Rahul
    Gupta, Deepak
    Khanna, Ashish
    Khamparia, Aditya
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2020, 16 (01)
  • [6] Local design for multi-layer perceptron
    Xu, Li
    Zidonghua Xuebao/Acta Automatica Sinica, 1997, 23 (03): : 325 - 331
  • [7] Symbolic representation of a multi-layer perceptron
    Mouria-Beji, F
    ARTIFICIAL NEURAL NETS AND GENETIC ALGORITHMS, 2001, : 205 - 208
  • [8] Graph Attention Multi-Layer Perceptron
    Zhang, Wentao
    Yin, Ziqi
    Sheng, Zeang
    Li, Yang
    Ouyang, Wen
    Li, Xiaosen
    Tao, Yangyu
    Yang, Zhi
    Cui, Bin
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 4560 - 4570
  • [9] The Application of Optimal Weights Initialization Algorithm Based on Information Amount in Multi-layer Perceptron Networks
    Xiao Wei
    Yan Xiu-tao
    PROCEEDINGS OF 2010 3RD IEEE INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND INFORMATION TECHNOLOGY (ICCSIT 2010), VOL 6, 2010, : 196 - 198
  • [10] Tighter Guarantees for the Compressive Multi-layer Perceptron
    Kaban, Ata
    Thummanusarn, Yamonporn
    THEORY AND PRACTICE OF NATURAL COMPUTING (TPNC 2018), 2018, 11324 : 388 - 400