Comparison of Deep-Learning-Based Segmentation Models: Using Top View Person Images

被引:50
作者
Ahmed, Imran [1 ]
Ahmad, Misbah [1 ]
Khan, Fakhri Alam [1 ]
Asif, Muhammad [2 ]
机构
[1] Inst Management Sci, Ctr Excellence IT, Peshawar 25000, Pakistan
[2] Natl Text Univ, Dept Comp Sci, Faisalabad 37610, Pakistan
关键词
Deep learning; semantic segmentation; top view person; FCN; U-Net; DeepLab; OBJECT DETECTION; DETECTOR;
D O I
10.1109/ACCESS.2020.3011406
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image segmentation is considered as a key research topic in the area of computer vision. It is pivotal in a broad range of real-life applications. Recently, the emergence of deep learning drives significant advancement in image segmentation; the developed systems are now capable of recognizing, segmenting, and classifying objects of specific interest in images. Generally, most of these techniques primarily focused on the asymmetric field of view or frontal view objects. This work explores widely used deep learning-based models for person segmentation using top view data set. The first model employed in this work is Fully Convolutional Neural Network (FCN) with Resnet-101 architecture. The network consists of a set of max-pooling and convolution layers to identify pixel-wise class labels and prediction of the mask. The second model is based on FCN called U-Net with Encoder-Decoder architecture. The encoder is mainly comprised of a contracting path, also called an encoder, which captures the context in the image and symmetric expanding path called decoder to enable accurate location. The third model used for top view person segmentation is a DeepLabV3 model also with encoder-decoder architecture. The encoder consists of trained Convolutional Neural Network (CNN) to encode feature maps of the input image. The decoder is used for up-sampling and reconstruction of output using important information extracted by the encoder. All segmentation models are firstly tested using pre-trained models (trained on frontal view data set). To improve the performance, these models are further trained using person data set captured from a top view. The output of all models consists of a segmented person in the top view images. The experimental results reveal the effectiveness and performance of segmentation models by achieving $IoU$ of 83%, 84%, and 86% and $mIoU$ of 80% 82% and 84% for FCN, U-Net, and DeepLabv3 respectively. Furthermore, the discussion is provided for output results with possible future guidelines.
引用
收藏
页码:136361 / 136373
页数:13
相关论文
共 60 条
[1]   Convolutional neural network-based person tracking using overhead views [J].
Ahmad, Misbah ;
Ahmed, Imran ;
Khan, Fakhri Alam ;
Qayum, Fawad ;
Aljuaid, Hanan .
INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2020, 16 (06)
[2]  
Ahmad M, 2018, 2018 9TH IEEE ANNUAL UBIQUITOUS COMPUTING, ELECTRONICS & MOBILE COMMUNICATION CONFERENCE (UEMCON), P746, DOI 10.1109/UEMCON.2018.8796595
[3]  
Ahmad M, 2019, INT J ADV COMPUT SC, V10, P567
[4]   Exploring Deep Learning Models for Overhead View Multiple Object Detection [J].
Ahmed, Imran ;
Din, Sadia ;
Jeon, Gwanggil ;
Piccialli, Francesco .
IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (07) :5737-5744
[5]   Person detector for different overhead views using machine learning [J].
Ahmed, Imran ;
Ahmad, Misbah ;
Adnan, Awais ;
Ahmad, Awais ;
Khan, Murad .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2019, 10 (10) :2657-2668
[6]   Efficient topview person detector using point based transformation and lookup table [J].
Ahmed, Imran ;
Ahmad, Misbah ;
Nawaz, Muhammad ;
Haseeb, Khalid ;
Khan, Sajidullah ;
Jeon, Gwanggil .
COMPUTER COMMUNICATIONS, 2019, 147 :188-197
[7]   A Robust Features-Based Person Tracker for Overhead Views in Industrial Environment [J].
Ahmed, Imran ;
Ahmad, Awais ;
Piccialli, Francesco ;
Sangaiah, Arun Kumar ;
Jeon, Gwanggil .
IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (03) :1598-1605
[8]  
[Anonymous], 2012, Advances in neural information processing systems, DOI DOI 10.5555/2999325.2999452
[9]  
[Anonymous], APPLICATIONS, V10, P300, DOI [10.14569/IJACSA.2019.0100339, DOI 10.14569/IJACSA.2021.0120957]
[10]  
[Anonymous], 2017, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2017.472