On Learning and Learned Data Representation by Capsule Networks

被引:6
|
作者
Lin, Ancheng [1 ]
Li, Jun [2 ,3 ]
Ma, Zhenyuan [4 ]
机构
[1] Guangdong Polytech Normal Univ, Sch Comp Sci, Guangzhou 510665, Guangdong, Peoples R China
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Sch Software, Sydney, NSW 2007, Australia
[3] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Artificial Intelligence, Sydney, NSW 2007, Australia
[4] Guangdong Polytech Normal Univ, Sch Math & Syst Sci, Guangzhou 510665, Guangdong, Peoples R China
关键词
Capsule network; deep neural network; interpretable learning; representation learning; RECOGNITION;
D O I
10.1109/ACCESS.2019.2911622
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Capsule networks (CapsNet) are recently proposed neural network models containing newly introduced processing layer, which are specialized in entity representation and discovery in images. CapsNet is motivated by a view of parse tree-like information processing mechanism and employs an iterative routing operation dynamically determining connections between layers composed of capsule units, in which the information ascends through different levels of interpretations, from raw sensory observation to semantically meaningful entities represented by active capsules. The CapsNet architecture is plausible and has been proven to be effective in some image data processing tasks, the newly introduced routing operation is mainly required for determining the capsules' activation status during the forward pass. However, its influence on model fitting and the resulted representation is barely understood. In this work, we investigate the following: 1) how the routing affects the CapsNet model fitting; 2) how the representation using capsules helps discover global structures in data distribution, and; 3) howthe learned data representation adapts and generalizes to new tasks. Our investigation yielded the results some of which have been mentioned in the original paper of CapsNet, they are: 1) the routing operation determines the certainty with which a layer of capsules pass information to the layer above and the appropriate level of certainty is related to the model fitness; 2) in a designed experiment using data with a known 2D structure, capsule representations enable a more meaningful 2D manifold embedding than neurons do in a standard convolutional neural network (CNN), and; 3) compared with neurons of the standard CNN, capsules of successive layers are less coupled and more adaptive to new data distribution.
引用
收藏
页码:50808 / 50822
页数:15
相关论文
共 50 条
  • [21] Data induced masking representation learning for face data analysis
    Guo, Tan
    Zhang, Lei
    Tan, Xiaoheng
    Yang, Liu
    Liang, Zhifang
    KNOWLEDGE-BASED SYSTEMS, 2019, 177 : 82 - 93
  • [22] Learning Joint Multimodal Representation with Adversarial Attention Networks
    Huang, Feiran
    Zhang, Xiaoming
    Li, Zhoujun
    PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 1874 - 1882
  • [23] A Comparative Study of Representation Learning Techniques for Dynamic Networks
    Vazquez, Carlos Ortega
    Mitrovic, Sandra
    De Weerdt, Jochen
    vanden Broucke, Seppe
    TRENDS AND INNOVATIONS IN INFORMATION SYSTEMS AND TECHNOLOGIES, VOL 3, 2020, 1161 : 523 - 530
  • [24] Learning Topological Representation for Networks via Hierarchical Sampling
    Fu, Guoji
    Hou, Chengbin
    Yao, Xin
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [25] Trembr: Exploring Road Networks for Trajectory Representation Learning
    Fu, Tao-Yang
    Lee, Wang-Chien
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2020, 11 (01)
  • [26] Scalable multi-modal representation learning networks
    Zihan Fang
    Ying Zou
    Shiyang Lan
    Shide Du
    Yanchao Tan
    Shiping Wang
    Artificial Intelligence Review, 58 (7)
  • [27] Multitask Representation Learning With Multiview Graph Convolutional Networks
    Huang, Hong
    Song, Yu
    Wu, Yao
    Shi, Jia
    Xie, Xia
    Jin, Hai
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (03) : 983 - 995
  • [28] Representation Learning Based on Path Selection in Complex Networks
    Liu Q.-X.
    Long H.
    Zheng P.-X.
    Long, Hang (2120161020@bit.edu.cn), 1600, Beijing Institute of Technology (40): : 282 - 289
  • [29] Hierarchical Prototype Networks for Continual Graph Representation Learning
    Zhang, Xikun
    Song, Dongjin
    Tao, Dacheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (04) : 4622 - 4636
  • [30] Representation Learning of Knowledge Graph for Wireless Communication Networks
    He, Shiwen
    Ou, Yeyu
    Wang, Liangpeng
    Zhan, Hang
    Ren, Peng
    Huang, Yongming
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 1338 - 1343