Multi-view representation learning for multi-view action recognition

被引:26
|
作者
Hao, Tong [1 ]
Wu, Dan [1 ]
Wang, Qian [1 ]
Sun, Jin-Sheng [1 ,2 ]
机构
[1] Tianjin Normal Univ, Tianjin Key Lab Anim & Plant Resistance, Coll Life Sci, Tianjin 300387, Peoples R China
[2] Tianjin Aquat Anim Infect Dis Control & Prevent C, Tianjin 300221, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-view learning; Multi-task learning; Sparse coding; Action recognition; MODEL; DICTIONARY;
D O I
10.1016/j.jvcir.2017.01.019
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Although multiple methods have been proposed for human action recognition, the existing multi-view approaches cannot well discover meaningful relationship among multiple action categories from different views. To handle this problem, this paper proposes an multi-view learning approach for multi-view action recognition. First, the proposed method leverages the popular visual representation method, bag of -visual-words (BoVW)/fisher vector (FV), to represent individual videos in each view. Second, the sparse coding algorithm is utilized to transfer the low-level features of various views into the discriminative and high-level semantics space. Third, we employ the multi-task learning (MTL) approach for joint action modeling and discovery of latent relationship among different action categories. The extensive experimental results on (MI)-I-2 and IXMAS datasets have demonstrated the effectiveness of our proposed approach. Moreover, the experiments further demonstrate that the discovered latent relationship can benefit multi-view model learning to augment the performance of action recognition. (C) 2017 Published by Elsevier Inc.
引用
收藏
页码:453 / 460
页数:8
相关论文
共 50 条
  • [21] Discriminative Multi-View Subspace Feature Learning for Action Recognition
    Sheng, Biyun
    Li, Jun
    Xiao, Fu
    Li, Qun
    Yang, Wankou
    Han, Junwei
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (12) : 4591 - 4600
  • [22] Multi-View Super Vector for Action Recognition
    Cai, Zhuowei
    Wang, Limin
    Peng, Xiaojiang
    Qiao, Yu
    2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, : 596 - 603
  • [23] Multi-view human action recognition: A survey
    Iosifidis, Alexandros
    Tefas, Anastasios
    Pitas, Ioannis
    2013 NINTH INTERNATIONAL CONFERENCE ON INTELLIGENT INFORMATION HIDING AND MULTIMEDIA SIGNAL PROCESSING (IIH-MSP 2013), 2013, : 522 - 525
  • [24] Generative Multi-View Human Action Recognition
    Wang, Lichen
    Ding, Zhengming
    Tao, Zhiqiang
    Liu, Yunyu
    Fu, Yun
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6221 - 6230
  • [25] Continuous Multi-View Human Action Recognition
    Wang, Qiang
    Sun, Gan
    Dong, Jiahua
    Wang, Qianqian
    Ding, Zhengming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (06) : 3603 - 3614
  • [26] View knowledge transfer network for multi-view action recognition
    Liang, Zixi
    Yin, Ming
    Gao, Junli
    He, Yicheng
    Huang, Weitian
    IMAGE AND VISION COMPUTING, 2022, 118
  • [27] DVANet: Disentangling View and Action Features for Multi-View Action Recognition
    Siddiqui, Nyle
    Tirupattur, Praveen
    Shah, Mubarak
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 5, 2024, : 4873 - 4881
  • [28] DEEP MULTI-VIEW ROBUST REPRESENTATION LEARNING
    Jiao, Zhenyu
    Xu, Chao
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 2851 - 2855
  • [29] Multi-View Concept Learning for Data Representation
    Guan, Ziyu
    Zhang, Lijun
    Peng, Jinye
    Fan, Jianping
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2015, 27 (11) : 3016 - 3028
  • [30] A survey on representation learning for multi-view data
    Qin, Yalan
    Zhang, Xinpeng
    Yu, Shui
    Feng, Guorui
    NEURAL NETWORKS, 2025, 181