On Learning and Learned Data Representation by Capsule Networks

被引:6
|
作者
Lin, Ancheng [1 ]
Li, Jun [2 ,3 ]
Ma, Zhenyuan [4 ]
机构
[1] Guangdong Polytech Normal Univ, Sch Comp Sci, Guangzhou 510665, Guangdong, Peoples R China
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Sch Software, Sydney, NSW 2007, Australia
[3] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Artificial Intelligence, Sydney, NSW 2007, Australia
[4] Guangdong Polytech Normal Univ, Sch Math & Syst Sci, Guangzhou 510665, Guangdong, Peoples R China
关键词
Capsule network; deep neural network; interpretable learning; representation learning; RECOGNITION;
D O I
10.1109/ACCESS.2019.2911622
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Capsule networks (CapsNet) are recently proposed neural network models containing newly introduced processing layer, which are specialized in entity representation and discovery in images. CapsNet is motivated by a view of parse tree-like information processing mechanism and employs an iterative routing operation dynamically determining connections between layers composed of capsule units, in which the information ascends through different levels of interpretations, from raw sensory observation to semantically meaningful entities represented by active capsules. The CapsNet architecture is plausible and has been proven to be effective in some image data processing tasks, the newly introduced routing operation is mainly required for determining the capsules' activation status during the forward pass. However, its influence on model fitting and the resulted representation is barely understood. In this work, we investigate the following: 1) how the routing affects the CapsNet model fitting; 2) how the representation using capsules helps discover global structures in data distribution, and; 3) howthe learned data representation adapts and generalizes to new tasks. Our investigation yielded the results some of which have been mentioned in the original paper of CapsNet, they are: 1) the routing operation determines the certainty with which a layer of capsules pass information to the layer above and the appropriate level of certainty is related to the model fitness; 2) in a designed experiment using data with a known 2D structure, capsule representations enable a more meaningful 2D manifold embedding than neurons do in a standard convolutional neural network (CNN), and; 3) compared with neurons of the standard CNN, capsules of successive layers are less coupled and more adaptive to new data distribution.
引用
收藏
页码:50808 / 50822
页数:15
相关论文
共 50 条
  • [31] Subdomain contraction in deep networks for robust representation learning
    Qi, Yu
    Pan, Zhentao
    Pan, Gang
    Wang, Yueming
    NEUROCOMPUTING, 2022, 513 : 318 - 328
  • [32] Co-Representation Learning Framework For the Open-Set Data Classification
    Wang, Zhuoyi
    Dong, Bo
    Lin, Yu
    Wang, Yigong
    Islam, Md Shihabul
    Khan, Latifur
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 239 - 244
  • [33] Capsule network-based data pruning in wireless sensor networks
    Umamaheswari, S.
    INTERNATIONAL JOURNAL OF COMMUNICATION SYSTEMS, 2020, 33 (05)
  • [34] Discrimination of Attention Deficit Hyperactivity Disorder Using Capsule Networks and LSTM Networks on fMRI Data
    Dey, Arunav
    Singh, Jigya
    Rathore, Manaswini
    Govind, Roshni
    Ladwani, Vandana M.
    24TH INTERNATIONAL CONFERENCE ON ENGINEERING APPLICATIONS OF NEURAL NETWORKS, EAAAI/EANN 2023, 2023, 1826 : 291 - 302
  • [35] FedClassAvg: Local Representation Learning for Personalized Federated Learning on Heterogeneous Neural Networks
    Jang, Jaehee
    Ha, Heonseok
    Jung, Dahuin
    Yoon, Sungroh
    51ST INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2022, 2022,
  • [36] A Data Adaptive Biological Sequence Representation for Supervised Learning
    Cakin H.
    Gorgulu B.
    Baydogan M.G.
    Zou N.
    Li J.
    Journal of Healthcare Informatics Research, 2018, 2 (4) : 448 - 471
  • [37] Multimodal Data Enhanced Representation Learning for Knowledge Graphs
    Wang, Zikang
    Li, Linjing
    Li, Qiudan
    Zeng, Daniel
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [38] Dense Representation Learning and Retrieval for Tabular Data Prediction
    Zheng, Lei
    Li, Ning
    Chen, Xianyu
    Gan, Quan
    Zhang, Weinan
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 3559 - 3569
  • [39] Representation Learning for the Clustering of Multi-Omics Data
    Viaud, Gautier
    Mayilvahanan, Prasanna
    Cournede, Paul-Henry
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2022, 19 (01) : 135 - 145
  • [40] Learning Interpretable Rules for Scalable Data Representation and Classification
    Wang, Zhuo
    Zhang, Wei
    Liu, Ning
    Wang, Jianyong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (02) : 1121 - 1133