3D-Pruning: A Model Compression Framework for Efficient 3D Action Recognition

被引:10
|
作者
Guo, Jinyang [1 ]
Liu, Jiaheng [2 ]
Xu, Dong [3 ]
机构
[1] Beihang Univ, Inst Artificial Intelligence, Beijing 100191, Peoples R China
[2] Beihang Univ, Sch Comp Sci & Engn, Beijing 100191, Peoples R China
[3] Univ Hong Kong, Dept Comp Sci, Pokfulam, Hong Kong, Peoples R China
关键词
Point cloud compression; Three-dimensional displays; Computational complexity; Computational modeling; Solid modeling; Task analysis; Feature extraction; Efficient deep learning; point cloud; 3D action recognition; model compression; OBJECT DETECTION; POINT; NETWORKS;
D O I
10.1109/TCSVT.2022.3197395
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The existing end-to-end optimized 3D action recognition methods often suffer from high computational costs. Observing that different frames and different points in point cloud sequences often have different importance values for the 3D action recognition task, in this work, we propose a fully automatic model compression framework called 3D-Pruning (3DP) for efficient 3D action recognition. After performing model compression by using our 3DP framework, the compressed model can process different frames and different points in each frame by using different computational complexities based on their importance values, in which both the importance value and computational complexity for each frame/point can be automatically learned. Extensive experiments on five benchmark datasets demonstrate the effectiveness of our 3DP framework for model compression.
引用
收藏
页码:8717 / 8729
页数:13
相关论文
共 50 条
  • [41] 3D Fingerprint Recognition based on Ridge-Valley-Guided 3D Reconstruction and 3D Topology Polymer Feature Extraction
    Yin, Xuefei
    Zhu, Yanming
    Hu, Jiankun
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (03) : 1085 - 1091
  • [42] Monocular 3D Object Detection Utilizing Auxiliary Learning With Deformable Convolution
    Chen, Jiun-Han
    Shieh, Jeng-Lun
    Haq, Muhamad Amirul
    Ruan, Shanq-Jang
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (03) : 2424 - 2436
  • [43] SCNet3D: Rethinking the Feature Extraction Process of Pillar-Based 3D Object Detection
    Li, Junru
    Wang, Zhiling
    Gong, Diancheng
    Wang, Chunchun
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2025, 26 (01) : 770 - 784
  • [44] Accurate and Real-Time 3D Pedestrian Detection Using an Efficient Attentive Pillar Network
    Le, Duy Tho
    Shi, Hengcan
    Rezatofighi, Hamid
    Cai, Jianfei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (02) : 1159 - 1166
  • [45] Action Recognition Using High Temporal Resolution 3D Neural Network Based on Dilated Convolution
    Xu, Yongyang
    Feng, Yaxing
    Xie, Zhong
    Xie, Mingyu
    Luo, Wei
    IEEE ACCESS, 2020, 8 : 165365 - 165372
  • [46] Multimodal Feature Fusion for 3D Shape Recognition and Retrieval
    Bu, Shuhui
    Cheng, Shaoguang
    Liu, Zhenbao
    Han, Junwei
    IEEE MULTIMEDIA, 2014, 21 (04) : 38 - 46
  • [47] MonoGRNet: A General Framework for Monocular 3D Object Detection
    Qin, Zengyi
    Wang, Jinglu
    Lu, Yan
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (09) : 5170 - 5184
  • [48] Efficient Spatio-Temporal Contrastive Learning for Skeleton-Based 3-D Action Recognition
    Gao, Xuehao
    Yang, Yang
    Zhang, Yimeng
    Li, Maosen
    Yu, Jin-Gang
    Du, Shaoyi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 405 - 417
  • [49] A Progressive Multi-View Learning Approach for Multi-Loss Optimization in 3D Object Recognition
    Prasad, Shitala
    Li, Yiqun
    Lin, Dongyun
    Dong, Sheng
    Nwe, Ma Tin Lay
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 707 - 711
  • [50] Merak: An Efficient Distributed DNN Training Framework With Automated 3D Parallelism for Giant Foundation Models
    Lai, Zhiquan
    Li, Shengwei
    Tang, Xudong
    Ge, Keshi
    Liu, Weijie
    Duan, Yabo
    Qiao, Linbo
    Li, Dongsheng
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (05) : 1466 - 1478