Appearance based pedestrians' gender recognition by employing stacked auto encoders in deep learning

被引:78
作者
Raza, Mudassar [1 ]
Sharif, Muhammad [1 ]
Yasmin, Mussarat [1 ]
Khan, Muhammad Attique [1 ]
Saba, Tanzila [2 ]
Fernandes, Steven Lawrence [3 ]
机构
[1] COMSATS Inst Informat Technol, Dept Comp Sci, Wah Cantt, Pakistan
[2] Prince Sultan Univ, Coll Comp & Informat Sci, Riyadh, Saudi Arabia
[3] Sahyadri Coll Engn & Management, Dept Elect & Commun Engn, Mangaluru, India
来源
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE | 2018年 / 88卷
关键词
Deep learning; Stacked sparse auto encoder; Pedestrians' gender classification; Pedestrian parsing; Soft max classifier; FACE RECOGNITION; NEURAL-NETWORK; CLASSIFICATION; SUPPORT; FUSION;
D O I
10.1016/j.future.2018.05.002
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Pedestrians' gender is a soft attribute which is useful in many areas of computer vision including human robot interaction, intelligent surveillance and human behavior analysis. Apart from its importance, pedestrians' gender prediction is one of the challenging methodologies in image processing. In this article, a deep learning approach is presented to classify a pedestrian as a male or a female. As a preprocessing step, pedestrian parsing is performed by a deep decompositional neural network method. The outcome of this network is a binary mask that maps the pedestrian full body from the input image. The pedestrian body image is then extracted by applying the generated pedestrian mask to the input image. This pre-processed image is then supplied to the stacked sparse auto encoder with soft max classifier for prediction. The proposed network is trained and tested separately on different pedestrians' views such as frontal views, back views and mixed views. The training is performed on PETA dataset. The experiments for testing are performed on MIT and PETA datasets (containing images other than train images). The accuracy values on MIT dataset are calculated as 82.9%, 81.8% and 82.4% on frontal, back and mixed views respectively. The mean AUC value by proposed scheme on PETA dataset is found as 91.5%+/- 4. The performance measures and comparisons with existing works depict the robustness and applicability of proposed methodology. (C) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:28 / 39
页数:12
相关论文
共 67 条
[1]   A distinctive approach in brain tumor detection and classification using MRI [J].
Amin, Javeria ;
Sharif, Muhammad ;
Yasmin, Mussarat ;
Fernandes, Steven Lawrence .
PATTERN RECOGNITION LETTERS, 2020, 139 :118-127
[2]   A method for the detection and classification of diabetic retinopathy using structural predictors of bright lesions [J].
Amin, Javeria ;
Sharif, Muhammad ;
Yasmin, Mussarat ;
Ali, Hussam ;
Fernandes, Steven Lawrence .
JOURNAL OF COMPUTATIONAL SCIENCE, 2017, 19 :153-164
[3]  
[Anonymous], IEEE ACCESS
[4]  
[Anonymous], 2017, ARXIV170202537
[5]  
[Anonymous], 2014, ARXIV14126885
[6]   Learned vs. Hand-Crafted Features for Pedestrian Gender Recognition [J].
Antipov, Grigory ;
Berrani, Sid-Ahmed ;
Ruchaud, Natacha ;
Dugelay, Jean-Luc .
MM'15: PROCEEDINGS OF THE 2015 ACM MULTIMEDIA CONFERENCE, 2015, :1263-1266
[7]   Minimalistic CNN-based ensemble model for gender prediction from face images [J].
Antipov, Grigory ;
Berrani, Sid-Ahmed ;
Dugelay, Jean-Luc .
PATTERN RECOGNITION LETTERS, 2016, 70 :59-65
[8]   Fundus Image Segmentation and Feature Extraction for the Detection of Glaucoma: A New Approach [J].
Bokhari, Syedia Tahseen Fatima ;
Sharif, Muhammad ;
Yasmin, Mussarat ;
Fernandes, Steven L. .
CURRENT MEDICAL IMAGING REVIEWS, 2018, 14 (01) :77-87
[9]  
Cao L., 2008, P 16 ACM INT C MULT, P725, DOI [10.1145/1459359.1459470, DOI 10.1145/1459359.1459470.11.P, DOI 10.1145/1459359.1459470]
[10]  
Castrillon-Santana M., 2015, PATTERN RECOGNIT LET