On Learning and Learned Data Representation by Capsule Networks

被引:6
|
作者
Lin, Ancheng [1 ]
Li, Jun [2 ,3 ]
Ma, Zhenyuan [4 ]
机构
[1] Guangdong Polytech Normal Univ, Sch Comp Sci, Guangzhou 510665, Guangdong, Peoples R China
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Sch Software, Sydney, NSW 2007, Australia
[3] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Artificial Intelligence, Sydney, NSW 2007, Australia
[4] Guangdong Polytech Normal Univ, Sch Math & Syst Sci, Guangzhou 510665, Guangdong, Peoples R China
关键词
Capsule network; deep neural network; interpretable learning; representation learning; RECOGNITION;
D O I
10.1109/ACCESS.2019.2911622
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Capsule networks (CapsNet) are recently proposed neural network models containing newly introduced processing layer, which are specialized in entity representation and discovery in images. CapsNet is motivated by a view of parse tree-like information processing mechanism and employs an iterative routing operation dynamically determining connections between layers composed of capsule units, in which the information ascends through different levels of interpretations, from raw sensory observation to semantically meaningful entities represented by active capsules. The CapsNet architecture is plausible and has been proven to be effective in some image data processing tasks, the newly introduced routing operation is mainly required for determining the capsules' activation status during the forward pass. However, its influence on model fitting and the resulted representation is barely understood. In this work, we investigate the following: 1) how the routing affects the CapsNet model fitting; 2) how the representation using capsules helps discover global structures in data distribution, and; 3) howthe learned data representation adapts and generalizes to new tasks. Our investigation yielded the results some of which have been mentioned in the original paper of CapsNet, they are: 1) the routing operation determines the certainty with which a layer of capsules pass information to the layer above and the appropriate level of certainty is related to the model fitness; 2) in a designed experiment using data with a known 2D structure, capsule representations enable a more meaningful 2D manifold embedding than neurons do in a standard convolutional neural network (CNN), and; 3) compared with neurons of the standard CNN, capsules of successive layers are less coupled and more adaptive to new data distribution.
引用
收藏
页码:50808 / 50822
页数:15
相关论文
共 50 条
  • [1] Representation Learning of Knowledge Graphs with Multi-scale Capsule Network
    Cheng, Jingwei
    Yang, Zhi
    Dang, Jinming
    Pan, Chunguang
    Zhang, Fu
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2019, PT I, 2019, 11871 : 282 - 290
  • [2] Imbalanced data learning by minority class augmentation using capsule adversarial networks
    Shamsolmoali, Pourya
    Zareapoor, Masoumeh
    Shen, Linlin
    Sadka, Abdul Hamid
    Yang, Jie
    NEUROCOMPUTING, 2021, 459 : 481 - 493
  • [3] On Representation Learning for Road Networks
    Wang, Meng-Xiang
    Lee, Wang-Chien
    Fu, Tao-Yang
    Yu, Ge
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2021, 12 (01)
  • [4] Contrastive representation learning for spectroscopy data analysis
    Vorozhtsov, Artem P.
    V. Kitina, Polina
    MENDELEEV COMMUNICATIONS, 2024, 34 (06) : 786 - 787
  • [5] Multi-Task Learning with Capsule Networks
    Lei, Kai
    Fu, Qiuai
    Liang, Yuzhi
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [6] Hyperbolic Representation Learning for Complex Networks
    Wang Q.
    Jiang H.
    Yi S.-W.
    Yang L.-T.
    Nai H.
    Nie Q.
    Ruan Jian Xue Bao/Journal of Software, 2021, 32 (01): : 93 - 117
  • [7] Multiple Kernel Representation Learning on Networks
    Celikkanat, Abdulkadir
    Shen, Yanning
    Malliaros, Fragkiskos D.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (06) : 6113 - 6125
  • [8] Contrastive representation learning on dynamic networks
    Jiao, Pengfei
    Chen, Hongjiang
    Tang, Huijun
    Bao, Qing
    Zhang, Long
    Zhao, Zhidong
    Wu, Huaming
    NEURAL NETWORKS, 2024, 174
  • [9] Object-centric Learning with Capsule Networks: A Survey
    Ribeiro, Fabio De Sousa
    Duarte, Kevin
    Everett, Miles
    Leontidis, Georgios
    Shah, Mubarak
    ACM COMPUTING SURVEYS, 2024, 56 (11)
  • [10] PhraseAttn: Dynamic Slot Capsule Networks for phrase representation in Neural Machine Translation
    Binh Nguyen
    Binh Le
    Nguyen, Long H. B.
    Dien Dinh
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 42 (04) : 3871 - 3878