Light Transport Induced Domain Adaptation for Semantic Segmentation in Thermal Infrared Urban Scenes

被引:17
作者
Chen, Junzhang [1 ]
Liu, Zichao [1 ]
Jin, Darui [1 ,2 ]
Wang, Yuanyuan [1 ]
Yang, Fan [1 ]
Bai, Xiangzhi [1 ,3 ,4 ]
机构
[1] Beijing Univ Aeronaut & Astronaut, Image Proc Ctr, Beijing 102206, Peoples R China
[2] Beihang Univ, Shen Yuan Honors Coll, Beijing 100191, Peoples R China
[3] Beihang Univ, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[4] Beihang Univ, Adv Innovat Ctr Biomed Engn, Beijing 100083, Peoples R China
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
Semantics; Image segmentation; Task analysis; Feature extraction; Cameras; Voltage control; Lighting; Urban scenes; thermal infrared image; domain adaptation; semantic segmentation; RECOGNITION;
D O I
10.1109/TITS.2022.3194931
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Semantic segmentation in urban scenes is widely used in applications of intelligent transportation systems (ITS). In urban scenes, thermal infrared (TIR) images can be captured in weak illumination conditions or in the presence of obscuration (e.g., light fog, smoke). Therefore, TIR images have great potential to endow automated intelligent vehicles or assist navigation systems. However, TIR imaging is blurry and low-contrast due to the absorption by atmospheric gases and heat transfer effect. Hence, TIR semantic segmentation in urban scenes has rarely been explored even though it has a wide range of scenarios in ITS. To overcome this limitation, we analyze the light transport of TIR light. Our analysis reveals that contours are the reliable features shared by TIR and Visible Spectrum (VS) light. Inspired by this, we attempt to transfer joint features from VS domain to TIR domain. Thus, we propose a curriculum domain adaptation method to guide the TIR urban scene semantic segmentation task from VS domain through contours. Moreover, to evaluate the proposed model, we build TIR-SS: an open-for-request dataset consisting of TIR images and pixel level annotations of 8 classes in urban scenes. Qualitative and quantitative experimental results on the dataset indicate that the proposed domain adaptation method outperforms related methods on this TIR semantic segmentation task.
引用
收藏
页码:23194 / 23211
页数:18
相关论文
共 100 条
  • [61] Speaker verification using adapted Gaussian mixture models
    Reynolds, DA
    Quatieri, TF
    Dunn, RB
    [J]. DIGITAL SIGNAL PROCESSING, 2000, 10 (1-3) : 19 - 41
  • [62] Romera E, 2019, IEEE INT VEH SYM, P1312, DOI [10.1109/IVS.2019.8813888, 10.1109/ivs.2019.8813888]
  • [63] ERFNet: Efficient Residual Factorized ConvNet for Real-Time Semantic Segmentation
    Romera, Eduardo
    Alvarez, Jose M.
    Bergasa, Luis M.
    Arroyo, Roberto
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2018, 19 (01) : 263 - 272
  • [64] U-Net: Convolutional Networks for Biomedical Image Segmentation
    Ronneberger, Olaf
    Fischer, Philipp
    Brox, Thomas
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION, PT III, 2015, 9351 : 234 - 241
  • [65] Adapting Visual Category Models to New Domains
    Saenko, Kate
    Kulis, Brian
    Fritz, Mario
    Darrell, Trevor
    [J]. COMPUTER VISION-ECCV 2010, PT IV, 2010, 6314 : 213 - +
  • [66] Maximum Classifier Discrepancy for Unsupervised Domain Adaptation
    Saito, Kuniaki
    Watanabe, Kohei
    Ushiku, Yoshitaka
    Harada, Tatsuya
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 3723 - 3732
  • [67] Sankaranarayanan, 2018, PROC IEEECVF C COMPU, P4085
  • [68] Generate To Adapt: Aligning Domains using Generative Adversarial Networks
    Sankaranarayanan, Swami
    Balaji, Yogesh
    Castillo, Carlos D.
    Chellappa, Rama
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 8503 - 8512
  • [69] Schroff F., 2008, P BRIT MACH VIS C, P1
  • [70] Generalized Domain-Adaptive Dictionaries
    Shekhar, Sumit
    Patel, Vishal M.
    Nguyen, Hien V.
    Chellappa, Rama
    [J]. 2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 361 - 368