DC3D: A Video Action Recognition Network Based on Dense Connection

被引:1
|
作者
Mu, Xiaofang [1 ]
Liu, Zhenyu [1 ]
Liu, Jiaji [1 ]
Li, Hao [1 ]
Li, Yue [2 ]
Li, Yikun [3 ]
机构
[1] Taiyuan Normal Univ, Coll Comp Sci & Technol, Taiyuan, Peoples R China
[2] Tianjin Univ Technol, Sch Elect & Informat Engn, Tianjin, Peoples R China
[3] Massey Univ, Dept Informat Sci, Palmerston North, New Zealand
来源
2022 TENTH INTERNATIONAL CONFERENCE ON ADVANCED CLOUD AND BIG DATA, CBD | 2022年
关键词
Action recognition; 3D Convolutions; DenseNet; Fisher discriminant criterion;
D O I
10.1109/CBD58033.2022.00032
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Efficiently extracting the temporal and spatial information of motion in the video, and how to obtain the spatiotemporal features with high degree of differentiation, is the key issue to improve the accuracy of action recognition classification. In this paper, a dense 3D convolutional block is designed as the basic unit to construct a dense convolutional 3D network, the spatiotemporal features existing in the video are extracted at the same time, and the transmission and reuse of the features in the network are strengthened, effectively fuse the shallow and deep spatiotemporal features of the network. At the same time, in order to make the features extracted by the network sufficiently discriminative, this paper proposes a joint loss function based on the Fisher discriminant regularization term, it can make the trained network have the ability to increase the inter-class dispersion and reduce the intra-class dispersion of the classified samples, and improve the classification accuracy. Experiments on the UCF-101 human actions classes dataset show that the network recognition accuracy rate proposed in this paper reaches 92.4%, which is higher than 85.2% of the C3D network, which proves the effectiveness of the method proposed in this paper.
引用
收藏
页码:133 / 138
页数:6
相关论文
共 50 条
  • [1] Dense Dilated Network for Video Action Recognition
    Xu, Baohan
    Ye, Hao
    Zheng, Yingbin
    Wang, Heng
    Luwang, Tianyu
    Jiang, Yu-Gang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (10) : 4941 - 4953
  • [2] Sparse Dense Transformer Network for Video Action Recognition
    Qu, Xiaochun
    Zhang, Zheyuan
    Xiao, Wei
    Ran, Jinye
    Wang, Guodong
    Zhang, Zili
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT II, 2022, 13369 : 43 - 56
  • [3] Spatiotemporal distilled dense-connectivity network for video action recognition
    Hao, Wangli
    Zhang, Zhaoxiang
    PATTERN RECOGNITION, 2019, 92 : 13 - 24
  • [4] Badminton video action recognition based on time network
    Zhi, Juncai
    Sun, Zijie
    Zhang, Ruijie
    Zhao, Zhouxiang
    JOURNAL OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING, 2023, 23 (05) : 2739 - 2752
  • [5] Human-Body Action Recognition Based on Dense Trajectories and Video Saliency
    Gao Deyong
    Kang Zibing
    Wang Song
    Wang Yangping
    LASER & OPTOELECTRONICS PROGRESS, 2020, 57 (24)
  • [6] Temporal Segment Connection Network for Action Recognition
    Li, Qian
    Yang, Wenzhu
    Chen, Xiangyang
    Yuan, Tongtong
    Wang, Yuxia
    IEEE ACCESS, 2020, 8 : 179118 - 179127
  • [7] Deep Spatiotemporal Relation Learning With 3D Multi-Level Dense Fusion for Video Action Recognition
    Zhang, Junxuan
    Hu, Haifeng
    IEEE ACCESS, 2019, 7 : 15222 - 15229
  • [8] Action Recognition by 3D Convolutional Network
    Brezovsky, Matus
    Sopiak, Dominik
    Oravec, Milos
    PROCEEDINGS OF ELMAR-2018: 60TH INTERNATIONAL SYMPOSIUM ELMAR-2018, 2018, : 71 - 74
  • [9] MNv3-MFAE: A Lightweight Network for Video Action Recognition
    Liu, Jie
    Liu, Wenyue
    Han, Ke
    ELECTRONICS, 2025, 14 (05):
  • [10] Dense Semantics-Assisted Networks for Video Action Recognition
    Luo, Haonan
    Lin, Guosheng
    Yao, Yazhou
    Tang, Zhenmin
    Wu, Qingyao
    Hua, Xiansheng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (05) : 3073 - 3084