DeepHSAR: Semi-supervised fine-grained learning for multi-label human sexual activity recognition

被引:1
作者
Gangwar, Abhishek [1 ,2 ]
Gonzalez-Castro, Victor [1 ]
Alegre, Enrique [1 ]
Fidalgo, Eduardo [1 ]
Martinez-Mendoza, Alicia [1 ]
机构
[1] Univ Leon, Dept Elect Engn & Automat Syst, Campus Vegazana S-N, E-24071 Leon, Spain
[2] Ctr Dev Adv Comp CDAC, Mumbai 400049, Maharashtra, India
关键词
Multi-label classification; Sexual activity detection; Fine-grained classification; Semi-supervised classification; Pornography detection; CONVOLUTIONAL NETWORKS; CLASSIFICATION; PORNOGRAPHY; ATTENTION; CONVNET;
D O I
10.1016/j.ipm.2024.103800
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The identification of sexual activities in images can be helpful in detecting the level of content severity and can assist pornography detectors in filtering specific types of content. In this paper, we propose a Deep Learning -based framework, named DeepHSAR, for semisupervised fine-grained multi -label Human Sexual Activity Recognition (HSAR). To the best of our knowledge, this is the first work to propose an approach to HSAR. We also introduce a new multi -label dataset, named SexualActs-150k, containing 150k images manually labeled with 19 types of sexual activities. DeepHSAR has two multi -label classification streams: one for global image representation and another for fine-grained representation. To perform finegrained image classification without ground -truth bounding box annotations, we propose a novel semi -supervised approach for multi -label fine-grained recognition, which learns through an iterative clustering and iterative CNN training process. We obtained a significant performance gain by fusing both streams (i.e., overall F1 -score of 79.29%), compared to when they work separately. The experiments demonstrate that the proposed framework explicitly outperforms baseline and state-of-the-art approaches. In addition, the proposed framework also obtains stateof-the-art or competitive results in semi -supervised multi -label learning experiments on the NUS -WIDE and MS-COCO datasets with overall F1 -scores of 75.98% and 85.17%, respectively. Furthermore, the proposed DeepHSAR has been assessed on the NPDI Pornography -2k video dataset, achieving a new state-of-the-art with 99.85% accuracy.
引用
收藏
页数:21
相关论文
共 59 条
[1]  
Chen SF, 2018, AAAI CONF ARTIF INTE, P6714
[2]   Knowledge-Guided Multi-Label Few-Shot Learning for General Image Recognition [J].
Chen, Tianshui ;
Lin, Liang ;
Chen, Riquan ;
Hui, Xiaolu ;
Wu, Hefeng .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (03) :1371-1384
[3]   Learning Semantic-Specific Graph Representation for Multi-Label Image Recognition [J].
Chen, Tianshui ;
Xu, Muxin ;
Hui, Xiaolu ;
Wu, Hefeng ;
Lin, Liang .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :522-531
[4]   BLOCK-DBSCAN: Fast clustering for large scale data [J].
Chen, Yewang ;
Zhou, Lida ;
Bouguila, Nizar ;
Wang, Cheng ;
Chen, Yi ;
Du, Jixiang .
PATTERN RECOGNITION, 2021, 109
[5]   Learning Graph Convolutional Networks for Multi-Label Recognition and Applications [J].
Chen, Zhao-Min ;
Wei, Xiu-Shen ;
Wang, Peng ;
Guo, Yanwen .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (06) :6969-6983
[6]   MULTI-LABEL IMAGE RECOGNITION WITH JOINT CLASS-AWARE MAP DISENTANGLING AND LABEL CORRELATION EMBEDDING [J].
Chen, Zhao-Min ;
Wei, Xiu-Shen ;
Jin, Xin ;
Guo, Yanwen .
2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, :622-627
[7]  
Cheng X., 2021, 2022 IEEE INT C MULT, P1
[8]  
Chua Tat-Seng, 2009, P ACM INT C IM VID R, P6
[9]   Learning a Deep ConvNet for Multi-label Classification with Partial Labels [J].
Durand, Thibaut ;
Mehrasa, Nazanin ;
Mori, Greg .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :647-657
[10]   AttM-CNN: Attention and metric learning based CNN for pornography, age and Child Sexual Abuse (CSA) Detection in images [J].
Gangwar, Abhishek ;
Gonzalez-Castro, Victor ;
Alegre, Enrique ;
Fidalgo, Eduardo .
NEUROCOMPUTING, 2021, 445 (445) :81-104