DeepHSAR: Semi-supervised fine-grained learning for multi-label human sexual activity recognition

被引:1
作者
Gangwar, Abhishek [1 ,2 ]
Gonzalez-Castro, Victor [1 ]
Alegre, Enrique [1 ]
Fidalgo, Eduardo [1 ]
Martinez-Mendoza, Alicia [1 ]
机构
[1] Univ Leon, Dept Elect Engn & Automat Syst, Campus Vegazana S-N, E-24071 Leon, Spain
[2] Ctr Dev Adv Comp CDAC, Mumbai 400049, Maharashtra, India
关键词
Multi-label classification; Sexual activity detection; Fine-grained classification; Semi-supervised classification; Pornography detection; CONVOLUTIONAL NETWORKS; CLASSIFICATION; PORNOGRAPHY; ATTENTION; CONVNET;
D O I
10.1016/j.ipm.2024.103800
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The identification of sexual activities in images can be helpful in detecting the level of content severity and can assist pornography detectors in filtering specific types of content. In this paper, we propose a Deep Learning -based framework, named DeepHSAR, for semisupervised fine-grained multi -label Human Sexual Activity Recognition (HSAR). To the best of our knowledge, this is the first work to propose an approach to HSAR. We also introduce a new multi -label dataset, named SexualActs-150k, containing 150k images manually labeled with 19 types of sexual activities. DeepHSAR has two multi -label classification streams: one for global image representation and another for fine-grained representation. To perform finegrained image classification without ground -truth bounding box annotations, we propose a novel semi -supervised approach for multi -label fine-grained recognition, which learns through an iterative clustering and iterative CNN training process. We obtained a significant performance gain by fusing both streams (i.e., overall F1 -score of 79.29%), compared to when they work separately. The experiments demonstrate that the proposed framework explicitly outperforms baseline and state-of-the-art approaches. In addition, the proposed framework also obtains stateof-the-art or competitive results in semi -supervised multi -label learning experiments on the NUS -WIDE and MS-COCO datasets with overall F1 -scores of 75.98% and 85.17%, respectively. Furthermore, the proposed DeepHSAR has been assessed on the NPDI Pornography -2k video dataset, achieving a new state-of-the-art with 99.85% accuracy.
引用
收藏
页数:21
相关论文
共 59 条
[51]   MIML-FCN plus : Multi-instance Multi-label Learning via Fully Convolutional Networks with Privileged Information [J].
Yang, Hao ;
Zhou, Joey Tianyi ;
Cai, Jianfei ;
Ong, Yew Soon .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :5996-6004
[52]  
Yazici VO, 2020, PROC CVPR IEEE, P13437, DOI 10.1109/CVPR42600.2020.01345
[53]   TransFGU: A Top-Down Approach to Fine-Grained Unsupervised Semantic Segmentation [J].
Yin, Zhaoyuan ;
Wang, Pichao ;
Wang, Fan ;
Xu, Xianzhe ;
Zhang, Hanling ;
Li, Hao ;
Jin, Rong .
COMPUTER VISION, ECCV 2022, PT XXIX, 2022, 13689 :73-89
[54]  
You RC, 2020, AAAI CONF ARTIF INTE, V34, P12709
[55]   Enhanced Adjacency Matrix-Based Lightweight Graph Convolution Network for Action Recognition [J].
Zhang, Daqing ;
Deng, Hongmin ;
Zhi, Yong .
SENSORS, 2023, 23 (14)
[56]   Multilabel Image Classification With Regional Latent Semantic Dependencies [J].
Zhang, Junjie ;
Wu, Qi ;
Shen, Chunhua ;
Zhang, Jian ;
Lu, Jianfeng .
IEEE TRANSACTIONS ON MULTIMEDIA, 2018, 20 (10) :2801-2813
[57]   Transformer-based Dual Relation Graph for Multi-label Image Recognition [J].
Zhao, Jiawei ;
Yan, Ke ;
Zhao, Yifan ;
Guo, Xiaowei ;
Huang, Feiyue ;
Li, Jia .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :163-172
[58]   Porn2Vec: A robust framework for detecting pornographic websites based on contrastive learning [J].
Zhao, Jun ;
Shao, Minglai ;
Peng, Hao ;
Wang, Hong ;
Li, Bo ;
Liu, Xudong .
KNOWLEDGE-BASED SYSTEMS, 2021, 228
[59]   Learning Spatial Regularization with Image-level Supervisions for Multi-label Image Classification [J].
Zhu, Feng ;
Li, Hongsheng ;
Ouyang, Wanli ;
Yu, Nenghai ;
Wang, Xiaogang .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2027-2036