Privacy-Safe Action Recognition via Cross-Modality Distillation

被引:0
|
作者
Kim, Yuhyun [1 ]
Jung, Jinwook [1 ]
Noh, Hyeoncheol [1 ]
Ahn, Byungtae [2 ]
Kwon, Junghye [3 ]
Choi, Dong-Geol [1 ]
机构
[1] Hanbat Natl Univ, Dept Informat & Commun Engn, Daejeon 34158, South Korea
[2] Korea Inst Machinery & Mat, Daejeon 34103, South Korea
[3] Chungnam Natl Univ, Coll Med, Dept Internal Med, Div Hematol Oncol, Daejeon 34134, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Action recognition; knowledge distillation; cross-modality distillation; deep learning; multi modal; privacy-safe;
D O I
10.1109/ACCESS.2024.3431227
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Human action recognition systems enhance public safety by detecting abnormal behavior autonomously. RGB sensors commonly used in such systems capture personal information of subjects and, as a result, run the risk of potential privacy leakage. On the other hand, privacy-safe alternatives, such as depth or thermal sensors, exhibit poorer performance because they lack the semantic context provided by RGB sensors. Moreover, the data availability of privacy-safe alternatives is significantly lower than RGB sensors. To address these problems, we explore effective cross-modality distillation methods in this paper, aiming to distill the knowledge of context-rich large-scale pre-trained RGB-based models into privacy-safe depth-based models. Based on extensive experiments on multiple architectures and benchmark datasets, we propose an effective method for training privacy-safe depth-based action recognition models via cross-modality distillation: cross-modality mixing distillation. This approach improves both the performance and efficiency by enabling interaction between depth and RGB modalities through a linear combination of their features. By utilizing the proposed cross-modal mixing distillation approach, we achieve state-of-the-art accuracy in two depth-based action recognition benchmarks. The code and the pre-trained models will be available upon publication.
引用
收藏
页码:125955 / 125965
页数:11
相关论文
empty
未找到相关数据