CrackFormer Network for Pavement Crack Segmentation

被引:39
作者
Liu, Huajun [1 ]
Yang, Jing [1 ]
Miao, Xiangyu [1 ]
Mertz, Christoph [2 ]
Kong, Hui [3 ,4 ,5 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[2] Carnegie Mellon Univ, Robot Inst, Pittsburgh, PA 15213 USA
[3] Univ Macau UM, State Key Lab Internet Things Smart City SKL IOTSC, Taipa, Macau, Peoples R China
[4] Univ Macau UM, Dept Electromech Engn EME, Taipa, Macau, Peoples R China
[5] Univ Macau UM, Dept Comp & Informat Sci CIS, Taipa, Macau, Peoples R China
关键词
Automatic crack segmentation; SegNet; ConvNet; transformer; CrackFormer; DAMAGE DETECTION;
D O I
10.1109/TITS.2023.3266776
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
In this paper, we rethink our earlier work on self-attention based crack segmentation, and propose an upgraded CrackFormer network (CrackFormer-II) for pavement crack segmentation, instead of only for fine-grained crackdetection tasks. This work embeds novel Transformer encoder modules into a SegNet-like encoder-decoder structure, where the basic module is composed of novel Transformer encoder blocks with effective relative positional embedding and long range interactions to extract efficient contextual information from feature-channels. Further, fusion modules of scaling-attention are proposed to integrate the results of each respective encoder and decoder block to highlight semantic features and suppress nonsemantic ones. Moreover, we update the Transformer encoder blocks enhanced by the local feed-forward layer and skipconnections, and optimize the channel configurations to compress the model parameters. Compared with the original CrackFormer, the CrackFormer-II is trained and evaluated on more general crack datasets. It achieves higher accuracy than the original CrackFormer, and the state-of-the-art (SOTA) method with 6.7x fewer FLOPs and 6.2x fewer parameters, and its practical inference speed is comparable to most classical CNN models. The experimental results show that it achieves the F-measures on Optimal Dataset Scale (ODS) of 0.912, 0.908, 0.914 and 0.869, respectively, on the four benchmarks. Codes are available at https://github.com/LouisNUST/CrackFormer-II.
引用
收藏
页码:9240 / 9252
页数:13
相关论文
共 52 条
  • [1] Image-based retrieval of concrete crack properties for bridge inspection
    Adhikari, R. S.
    Moselhi, O.
    Bagchi, A.
    [J]. AUTOMATION IN CONSTRUCTION, 2014, 39 : 180 - 194
  • [2] Alfarrarjeh A, 2018, IEEE INT CONF BIG DA, P5201, DOI 10.1109/BigData.2018.8621899
  • [3] Amhaz R, 2014, IEEE IMAGE PROC, P788, DOI 10.1109/ICIP.2014.7025158
  • [4] SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation
    Badrinarayanan, Vijay
    Kendall, Alex
    Cipolla, Roberto
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) : 2481 - 2495
  • [5] Bello I., 2021, PROC INT C LEARN REP, P1
  • [6] Attention Augmented Convolutional Networks
    Bello, Irwan
    Zoph, Barret
    Vaswani, Ashish
    Shlens, Jonathon
    Le, Quoc V.
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 3285 - 3294
  • [7] Deep Learning-Based Crack Damage Detection Using Convolutional Neural Networks
    Cha, Young-Jin
    Choi, Wooram
    Buyukozturk, Oral
    [J]. COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2017, 32 (05) : 361 - 378
  • [8] Twin-field quantum key distribution over a 511km optical fibre linking two distant metropolitan areas
    Chen, Jiu-Peng
    Zhang, Chi
    Liu, Yang
    Jiang, Cong
    Zhang, Wei-Jun
    Han, Zhi-Yong
    Ma, Shi-Zhao
    Hu, Xiao-Long
    Li, Yu-Huai
    Liu, Hui
    Zhou, Fei
    Jiang, Hai-Feng
    Chen, Teng-Yun
    Li, Hao
    You, Li-Xing
    Wang, Zhen
    Wang, Xiang-Bin
    Zhang, Qiang
    Pan, Jian-Wei
    [J]. NATURE PHOTONICS, 2021, 15 (08) : 570 - 575
  • [9] Fast Edge Detection Using Structured Forests
    Dollar, Piotr
    Zitnick, C. Lawrence
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2015, 37 (08) : 1558 - 1570
  • [10] Dong YH, 2021, PR MACH LEARN RES, V139