Potential Layer-Wise Supervised Learning for Training Multi-Layered Neural Networks

被引:0
作者
Kamimura, Ryotaro [1 ,2 ]
机构
[1] Tokai Univ, IT Educ Ctr, 4-1-1 Kitakaname, Hiratsuka, Kanagawa 2591292, Japan
[2] Tokai Univ, Grad Sch Sci & Technol, 4-1-1 Kitakaname, Hiratsuka, Kanagawa 2591292, Japan
来源
2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2017年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The present paper tries to show that the greedy layer-wise supervised learning becomes effective enough to improve generalization and interpretation by the help of potential learning. It has been observed that unsupervised pre-training has a shortcoming of vanishing information as is the case of simple multi-layered network training. When the layer becomes higher, valuable information becomes smaller. Information through many different layers tends to diminish considerably and naturally from the information-theoretic point of view. For this, we use the layer-wise supervised training to prevent information from diminishing. The supervised learning has been said to be not good for pre-training for multi-layered neural networks. However, we have found that the new potential learning can be effectively used to extract valuable information through supervised pre-training. With the help of important components extracted by the potential learning, the supervised pre-training becomes effective for training multi-layered neural networks. We applied the method to two data sets, namely, an artificial and banknote data sets. In both cases, the potential learning proved to be effective in increasing generalization performance. In addition, we could show a possibility that final representation by this method could be clearly understood.
引用
收藏
页码:2568 / 2575
页数:8
相关论文
共 16 条
  • [1] [Anonymous], 2006, NIPS
  • [2] [Anonymous], 1963, Information Theory and Coding
  • [3] ATICK JJ, 1992, NETWORK-COMP NEURAL, V3, P213, DOI [10.1088/0954-898X/3/2/009, 10.3109/0954898X.2011.638888]
  • [4] Bache K., 2013, UCI Machine Learning Repository
  • [5] Learning Deep Architectures for AI
    Bengio, Yoshua
    [J]. FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2009, 2 (01): : 1 - 127
  • [6] Cover TM., 1991, ELEMENTS INFORM THEO, V1, P279
  • [7] Deco G., 2012, An Information-Theoretic Approach to Neural Computing
  • [8] Glorot X., 2011, P 14 INT C ARTIFICIA, P315
  • [9] Reducing the dimensionality of data with neural networks
    Hinton, G. E.
    Salakhutdinov, R. R.
    [J]. SCIENCE, 2006, 313 (5786) : 504 - 507
  • [10] Kamimura R., 2017, SELECTIVE COOPERATIV