U-ASD Net: Supervised Crowd Counting Based on Semantic Segmentation and Adaptive Scenario Discovery

被引:10
作者
Hafeezallah, Adel [1 ]
Al-Dhamari, Ahlam [2 ,3 ]
Abu-Bakar, Syed Abd Rahman [3 ]
机构
[1] Taibah Univ, Dept Elect Engn, Madinah, Saudi Arabia
[2] Univ Teknol Malaysia, Fac Engn, Sch Elect Engn, Dept Elect & Comp Engn, Johor Baharu 81310, Malaysia
[3] Hodeidah Univ, Dept Comp Engn, Hodeidah, Yemen
来源
IEEE ACCESS | 2021年 / 9卷
关键词
Computer vision; deep learning; crowd counting; density map estimation; U-Net; adaptive scenario discovery; NETWORK;
D O I
10.1109/ACCESS.2021.3112174
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Crowd counting considers one of the most significant and challenging issues in computer vision and deep learning communities, whose applications are being utilized for various tasks. While this issue is well studied, it remains an open challenge to manage perspective distortions and scale variations. How well these problems are resolved has a huge impact on predicting a high-quality crowd density map. In this study, a hybrid and modified deep neural network (U-ASD Net), based on U-Net and adaptive scenario discovery (ASD), is proposed to get precise and effective crowd counting. The U part is produced by replacing the nearest upsampling in the encoder of U-Net with max-unpooling. This modification provides a better crowd counting performance by capturing more spatial information. The max-unpooling layers upsample the feature maps based on the max locations held from the downsampling process. The ASD part is constructed with three light pathways, two of which have been learned to reflect various densities of the crowd and define the appropriate geometric configuration employing various sizes of the receptive field. The third pathway is an adaptation path, which implicitly discovers and models complex scenarios to recalibrate pathway-wise responses adaptively. ASD has no additional branches to avoid increasing the complexity. The designed model is end-to-end trainable. This integration provides an effective model to count crowds in both dense and sparse datasets. It also predicts an elevated quality density map with a high structural similarity index and a high peak signal-to-noise ratio. Several comprehensive experiments on four popular datasets for crowd counting have been carried out to demonstrate the proposed method's promising performance compared to other state-of-the-art approaches. Furthermore, a new dataset with its manual annotations, called Haramain with three different scenes and different densities, is introduced and used for evaluating the U-ASD Net.
引用
收藏
页码:127444 / 127459
页数:16
相关论文
共 76 条
  • [1] Al-Dhamari A., 2019, TELKOMNIKA (Telecommun. Comput. Electron. Control), V17, P2039, DOI DOI 10.12928/TELKOMNIKA.V17I4.12753
  • [2] Al-Dhamari A. K., 2017, J SOFTWARE ENG APPL, V10, P56, DOI DOI 10.4236/JSEA.2017.101004
  • [3] Abnormal behavior detection using sparse representations through sequential generalization of k-means
    Al-Dhamari, Ahlam
    Sudirman, Rubita
    Mahmood, Nasrul Humaimi
    [J]. TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2021, 29 (01) : 152 - 168
  • [4] Transfer Deep Learning Along With Binary Support Vector Machine for Abnormal Behavior Detection
    Al-Dhamari, Ahlam
    Sudirman, Rubita
    Mahmood, Nasrul Humaimi
    [J]. IEEE ACCESS, 2020, 8 : 61085 - 61095
  • [5] Allah AAH, 2014, 2014 IEEE INTERNATIONAL CONFERENCE ON CONTROL SYSTEM COMPUTING AND ENGINEERING, P502, DOI 10.1109/ICCSCE.2014.7072770
  • [6] [Anonymous], ARXIV170309393
  • [7] Cao X., 2018, P EUR C COMP VIS ECC, P734
  • [8] Privacy preserving crowd monitoring: Counting people without people models or tracking
    Chan, Antoni B.
    Liang, Zhang-Sheng John
    Vasconcelos, Nuno
    [J]. 2008 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-12, 2008, : 1766 - 1772
  • [9] Chen J., NEUROCOMPUTING, V420, P82
  • [10] Feature Mining for Localised Crowd Counting
    Chen, Ke
    Loy, Chen Change
    Gong, Shaogang
    Xiang, Tao
    [J]. PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2012, 2012,