J-LDFR: joint low-level and deep neural network feature representations for pedestrian gender classification

被引:19
作者
Fayyaz, Muhammad [1 ]
Yasmin, Mussarat [1 ]
Sharif, Muhammad [1 ]
Raza, Mudassar [1 ]
机构
[1] COMSATS Univ Islamabad, Dept Comp Sci, Wah Campus, Wah Cantt, Pakistan
关键词
Handcrafted features; Deep learning; Joint feature representations; Visual surveillance applications; LOCAL BINARY PATTERNS; RECOGNITION; FUSION; HYBRID; ALGORITHM; OBJECT; AGE;
D O I
10.1007/s00521-020-05015-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Appearance-based gender classification is one of the key areas in pedestrian analysis, and it has many useful applications such as visual surveillance, predict demographics statistics, population prediction, and human-computer interaction. For pedestrian gender classification, traditional and deep convolutional neural network (CNN) approaches are employed individually. However, they are facing issues, for instance, discriminative feature representations, lower classification accuracy, and small sample size for model learning. To address these issues, this article proposes a framework that considers the combination of both traditional and deep CNN approaches for gender classification. To realize it, HOG- and LOMO-assisted low-level features are extracted to handle rotation, viewpoint and illumination variances in the images. Simultaneously, VGG19- and ResNet101-based standard deep CNN architectures are employed to acquire the deep features which are robust against pose variations. To avoid the ambiguous and unnecessary feature representations, the entropy-controlled features are picked from both low-level and deep representations of features that reduce the dimension of computed features. By merging the selected low-level features with deep features, we obtain a robust joint feature representation. The extensive experiments are conducted on PETA and MIT datasets, and computed results suggest that using the integration of both low-level and deep feature representations can improve the performance as compared to using these feature representations, individually. The proposed framework achieves AU-ROC of 96% and accuracy of 89.3% on the PETA dataset, and AU-ROC of 86% and accuracy of 82% on the MIT dataset. The experimental outcomes show that the proposed J-LDFR framework outperformed the existing gender classification methods.
引用
收藏
页码:361 / 391
页数:31
相关论文
共 102 条
[1]   AFIF4: Deep gender classification based on AdaBoost-based fusion of isolated facial features and foggy faces [J].
Afifi, Mahmoud ;
Abdelhamed, Abdelrahman .
JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2019, 62 :77-86
[2]  
Ahmad K, 2018, INT ARAB CONF INF TE, P273
[3]  
Ahmed HA, 2016, INT J ADV COMPUT SC, V7, P101
[4]  
Ahmed SS., 2019, UKH J SCI ENG, V3, P31, DOI [10.25079/ukhjse.v3n2y2019.pp31-4, DOI 10.25079/UKHJSE.V3N2Y2019.PP31-40]
[5]   Gender recognition: A multiscale decision fusion approach [J].
Alexandre, Luis A. .
PATTERN RECOGNITION LETTERS, 2010, 31 (11) :1422-1427
[6]   Kernel Visual Keyword Description for Object and Place Recognition [J].
Ali, Abbas M. ;
Rashid, Tarik A. .
ADVANCES IN SIGNAL PROCESSING AND INTELLIGENT RECOGNITION SYSTEMS (SIRS-2015), 2016, 425 :27-38
[7]  
Amayeh G., 2008, 2008 IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. Work. CVPR Work, DOI [10.1109/CVPRW.2008.4563122, DOI 10.1109/CVPRW.2008.4563122]
[8]   Learned vs. Hand-Crafted Features for Pedestrian Gender Recognition [J].
Antipov, Grigory ;
Berrani, Sid-Ahmed ;
Ruchaud, Natacha ;
Dugelay, Jean-Luc .
MM'15: PROCEEDINGS OF THE 2015 ACM MULTIMEDIA CONFERENCE, 2015, :1263-1266
[9]  
ARORA S, 2018, 9 INT C COMP COMM NE, P1
[10]   Comparative Study of Gait Gender Identification using Gait Energy Image (GEI) and Gait Information Image (GII) [J].
Asmara, Rosa Andrie ;
Masruri, Irtafa ;
Rahmad, Cahya ;
Siradjuddin, Indrazno ;
Rohadi, Erfan ;
Ronilaya, Ferdian ;
Handayani, Anik Nur ;
Hasanah, Qonitatul .
3RD ANNUAL APPLIED SCIENCE AND ENGINEERING CONFERENCE (AASEC 2018), 2018, 197