On Learning and Learned Data Representation by Capsule Networks

被引:6
|
作者
Lin, Ancheng [1 ]
Li, Jun [2 ,3 ]
Ma, Zhenyuan [4 ]
机构
[1] Guangdong Polytech Normal Univ, Sch Comp Sci, Guangzhou 510665, Guangdong, Peoples R China
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Sch Software, Sydney, NSW 2007, Australia
[3] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Artificial Intelligence, Sydney, NSW 2007, Australia
[4] Guangdong Polytech Normal Univ, Sch Math & Syst Sci, Guangzhou 510665, Guangdong, Peoples R China
关键词
Capsule network; deep neural network; interpretable learning; representation learning; RECOGNITION;
D O I
10.1109/ACCESS.2019.2911622
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Capsule networks (CapsNet) are recently proposed neural network models containing newly introduced processing layer, which are specialized in entity representation and discovery in images. CapsNet is motivated by a view of parse tree-like information processing mechanism and employs an iterative routing operation dynamically determining connections between layers composed of capsule units, in which the information ascends through different levels of interpretations, from raw sensory observation to semantically meaningful entities represented by active capsules. The CapsNet architecture is plausible and has been proven to be effective in some image data processing tasks, the newly introduced routing operation is mainly required for determining the capsules' activation status during the forward pass. However, its influence on model fitting and the resulted representation is barely understood. In this work, we investigate the following: 1) how the routing affects the CapsNet model fitting; 2) how the representation using capsules helps discover global structures in data distribution, and; 3) howthe learned data representation adapts and generalizes to new tasks. Our investigation yielded the results some of which have been mentioned in the original paper of CapsNet, they are: 1) the routing operation determines the certainty with which a layer of capsules pass information to the layer above and the appropriate level of certainty is related to the model fitness; 2) in a designed experiment using data with a known 2D structure, capsule representations enable a more meaningful 2D manifold embedding than neurons do in a standard convolutional neural network (CNN), and; 3) compared with neurons of the standard CNN, capsules of successive layers are less coupled and more adaptive to new data distribution.
引用
收藏
页码:50808 / 50822
页数:15
相关论文
共 50 条
  • [41] Capsule Attention for Multimodal EEG-EOG Representation Learning With Application to Driver Vigilance Estimation
    Zhang, Guangyi
    Etemad, Ali
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2021, 29 : 1138 - 1149
  • [42] SHNE: Representation Learning for Semantic-Associated Heterogeneous Networks
    Zhang, Chuxu
    Swami, Ananthram
    Chawla, Nitesh, V
    PROCEEDINGS OF THE TWELFTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM'19), 2019, : 690 - 698
  • [43] Biologically Plausible Learning of Text Representation with Spiking Neural Networks
    Bialas, Marcin
    Mironczuk, Marcin Michal
    Mandziuk, Jacek
    PARALLEL PROBLEM SOLVING FROM NATURE - PPSN XVI, PT I, 2020, 12269 : 433 - 447
  • [44] Representation Learning for Classification in Heterogeneous Graphs with Application to Social Networks
    Dos Santos, Ludovic
    Piwowarski, Benjamin
    Denoyer, Ludovic
    Gallinari, Patrick
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2018, 12 (05)
  • [45] RLIM: representation learning method for influence maximization in social networks
    Chengai Sun
    Xiuliang Duan
    Liqing Qiu
    Qiang Shi
    Tengteng Li
    International Journal of Machine Learning and Cybernetics, 2022, 13 : 3425 - 3440
  • [46] Dynamic Representation Learning via Recurrent Graph Neural Networks
    Zhang, Chun-Yang
    Yao, Zhi-Liang
    Yao, Hong-Yu
    Huang, Feng
    Chen, C. L. Philip
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (02): : 1284 - 1297
  • [47] Generative Adversarial Networks for Multimodal Representation Learning in Video Hyperlinking
    Vukotic, Vedran
    Raymond, Christian
    Gravier, Guillaume
    PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR'17), 2017, : 421 - 424
  • [48] Intrusion Detection Using Convolutional Neural Networks for Representation Learning
    Li, Zhipeng
    Qin, Zheng
    Huang, Kai
    Yang, Xiao
    Ye, Shuxiong
    NEURAL INFORMATION PROCESSING, ICONIP 2017, PT V, 2017, 10638 : 858 - 866
  • [49] Structural Representation Learning for User Alignment Across Social Networks
    Liu, Li
    Li, Xin
    Cheung, William K.
    Liao, Lejian
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2020, 32 (09) : 1824 - 1837
  • [50] RLIM: representation learning method for influence maximization in social networks
    Sun, Chengai
    Duan, Xiuliang
    Qiu, Liqing
    Shi, Qiang
    Li, Tengteng
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2022, 13 (11) : 3425 - 3440