Real-time construction demolition waste detection using state-of-the-art deep learning methods; single-stage vs two-stage detectors

被引:36
作者
Demetriou, Demetris [1 ]
Mavromatidis, Pavlos [2 ]
Robert, Ponsian M. [2 ]
Papadopoulos, Harris [2 ]
Petrou, Michael F. [1 ]
Nicolaides, Demetris [2 ,3 ]
机构
[1] Univ Cyprus, Dept Civil & Environm Engn, CY-1303 Nicosia, Cyprus
[2] Frederick Res Ctr, CY-1036 Nicosia, Cyprus
[3] Frederick Univ, CY-1036 Nicosia, Cyprus
关键词
Construction and Demolition Waste; Object detection; Waste sorting; Deep learning; Convolutional neural networks; NEURAL-NETWORK;
D O I
10.1016/j.wasman.2023.05.039
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Central to the development of a successful waste sorting robot lies an accurate and fast object detection system. This study assesses the performance of the most representative deep-learning models for the real-time localisation and classification of Construction and Demolition Waste (CDW). For the investigation, both single-stage (SSD, YOLO) and two-stage (Faster-RCNN) detector architectures coupled with various backbone feature extractors (ResNet, MobileNetV2, efficientDet) were considered. A total of 18 models of variable depth were trained and tested on the first openly accessible CDW dataset developed by the authors of this study. This dataset consists of images of 6600 samples of CDW belonging to three object categories: brick, concrete, and tile. For an in-depth examination of the performance of the developed models under working conditions, two testing datasets containing normally and heavily stacked and adhered samples of CDW were developed. A comprehensive comparison between the different models yields that the latest version of the YOLO series (YoloV7) attains the best accuracy (mAP50:95 & AP;70%) at the highest inference speed (<30 ms), while also exhibiting enough precision to deal with severely stacked and adhered samples of CDW. Additionally, it was observed that despite the rising popularity of single-stage detectors, apart from YoloV7, Faster-RCNN models remain the most robust in terms of exhibiting the least mAP fluctuations over the testing datasets considered.
引用
收藏
页码:194 / 203
页数:10
相关论文
共 53 条
  • [11] Garcia M.C., 2021, REMOTE SENS BASEL, P13
  • [12] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [13] Densely Connected Convolutional Networks
    Huang, Gao
    Liu, Zhuang
    van der Maaten, Laurens
    Weinberger, Kilian Q.
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2261 - 2269
  • [14] Huang J., 2020, TENSORFLOW OBJECT DE
  • [15] Jocher Glenn, 2021, Zenodo
  • [16] Kretz D., 2018, DEV IMPLEMENTATION I
  • [17] Deep learning of grasping detection for a robot used in sorting construction and demolition waste
    Ku, Yuedong
    Yang, Jianhong
    Fang, Huaiying
    Xiao, Wen
    Zhuang, Jiangteng
    [J]. JOURNAL OF MATERIAL CYCLES AND WASTE MANAGEMENT, 2021, 23 (01) : 84 - 95
  • [18] Li CY, 2022, Arxiv, DOI [arXiv:2209.02976, 10.48550/arXiv.2209.02976, DOI 10.48550/ARXIV.2209.02976]
  • [19] Li J., 2022, WASTE MANAGE, V139
  • [20] Lin TY, 2018, Arxiv, DOI [arXiv:1708.02002, DOI 10.48550/ARXIV.1708.02002]