Efficient Dual-Stream Fusion Network for Real-Time Railway Scene Understanding

被引:1
作者
Cao, Zhiwei [1 ,2 ,3 ]
Gao, Yang [1 ,2 ,3 ]
Bai, Jie [1 ,2 ,3 ]
Qin, Yong [1 ,2 ,3 ]
Zheng, Yuanjin [4 ]
Jia, Limin [1 ,2 ,3 ]
机构
[1] Beijing Jiaotong Univ, State Key Lab Adv Rail Autonomous Operat, Beijing 100044, Peoples R China
[2] Beijing Jiaotong Univ, Key Lab Railway Ind Proact Safety & Risk Control, Beijing 100044, Peoples R China
[3] Beijing Jiaotong Univ, Sch Traff & Transportat, Beijing 100044, Peoples R China
[4] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
关键词
Rail transportation; Feature extraction; Semantic segmentation; Semantics; Safety; Rails; Real-time systems; Railway scene understanding; semantic segmentation; dual-stream fusion network; intelligent railway; railway safety; autonomous driving;
D O I
10.1109/TITS.2024.3377187
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Railway scene understanding is key to autonomous train operation and important in active train perception. However, most railway scene understanding methods focus on track extraction and ignore other components of railway scenes. Although several semantic segmentation algorithms are used to identify railway scenes, they are computationally expensive and slow with limits applications in railways. To solve these problems, we propose efficient dual-stream fusion network (EDFNet), a lightweight semantic segmentation algorithm, for understanding railway scenes. First, a dual-stream backbone network based on mobile inverted residual blocks is proposed to extract and fuse detailed features and semantic features. Next, a bi-directional feature pyramid pooling module is proposed to obtain multi-scale features and deep semantic features. Finally, a multi-task aggregate loss is designed to learn semantic and boundary information, thus improving the accuracy without increasing the computational complexity. Extensive experimental results demonstrate that EDFNet outperforms the lightweight state-of-the-art algorithms with high accuracy and fast speed on two railway datasets.
引用
收藏
页码:9442 / 9452
页数:11
相关论文
共 57 条
  • [21] NDNet: Spacewise Multiscale Representation Learning via Neighbor Decoupling for Real-Time Driving Scene Parsing
    Li, Shu
    Yan, Qingqing
    Zhou, Xun
    Wang, Deming
    Liu, Chengju
    Chen, Qijun
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (06) : 7884 - 7898
  • [22] Vanishing Point Detection and Rail Segmentation Based on Deep Multi-Task Learning
    Li, Xingxin
    Zhu, Liqiang
    Yu, Zujun
    Guo, Baoqing
    Wan, Yanqin
    [J]. IEEE ACCESS, 2020, 8 : 163015 - 163025
  • [23] Li Y., 2021, P 5 INT C EL ENG INF, P388
  • [24] Graph-guided Architecture Search for Real-time Semantic Segmentation
    Lin, Peiwen
    Sun, Peng
    Cheng, Guangliang
    Xie, Sirui
    Li, Xi
    Shi, Jianping
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 4202 - 4211
  • [25] Liu RP, 2024, Arxiv, DOI arXiv:2202.13393
  • [26] Structured Knowledge Distillation for Semantic Segmentation
    Liu, Yifan
    Chen, Ke
    Liu, Chris
    Qin, Zengchang
    Luo, Zhenbo
    Wang, Jingdong
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2599 - 2608
  • [27] A Vision-Based Approach for Rail Extraction and its Application in a Camera Pan-Tilt Control System
    Nassu, Bogdan Tomoyuki
    Ukai, Masato
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2012, 13 (04) : 1763 - 1771
  • [28] Deep Dual-Resolution Networks for Real-Time and Accurate Semantic Segmentation of Traffic Scenes
    Pan, Huihui
    Hong, Yuanduo
    Sun, Weichao
    Jia, Yisong
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (03) : 3448 - 3460
  • [29] Paszke A, 2016, Arxiv, DOI [arXiv:1606.02147, 10.48550/arXiv.1606.02147]
  • [30] Full-Resolution Residual Networks for Semantic Segmentation in Street Scenes
    Pohlen, Tobias
    Hermans, Alexander
    Mathias, Markus
    Leibe, Bastian
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3309 - 3318