Action-State Joint Learning-Based Vehicle Taillight Recognition in Diverse Actual Traffic Scenes

被引:8
作者
Song, Wenjie [1 ]
Liu, Shixian [1 ]
Zhang, Ting [1 ]
Yang, Yi [1 ]
Fu, Mengyin [1 ,2 ]
机构
[1] Beijing Inst Technol, Sch Automat, Beijing 100081, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Automat, Nanjing 210014, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Brakes; Time series analysis; Image recognition; Brightness; Hidden Markov models; Adaptation models; Vehicle taillight recognition; autonomous driving; behavior understanding; urban traffic;
D O I
10.1109/TITS.2022.3160501
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
As the vital factor of vehicle behavior understanding and prediction, vehicle taillight recognition is an important technology for autonomous driving, especially in diverse actual traffic scenes full of dynamic interactive traffic participants. However, in practical application, it always faces many challenges, such as `variable lighting conditions', `non-uniform taillight standards' and `random relative observation pose', which lead to few mature solutions in current common autopilot systems. This work proposes an action-state joint learning-based vehicle taillight recognition method on the basis of vehicles detection and tracking, which takes both taillight state features and time series features into account, consequently getting practicable results even in complex actual scenes. In detail, vehicle tracking sequence is used as input and split into pieces through a sliding window. Then, a CNN-LSTM model is applied to simultaneously identify the action features of brake lights and turn signals, dividing taillight actions into five categories: None, Brake_on, Brake_off, Left_turn, Right_turn. Next, the brightness of high-position brake light is extracted through semantic segmentation and combined with taillight actions to form higher-level features for taillight state sequence analysis. Finally, an undirected graph model is used to establish the long-term dependence between successive pieces by analysing the higher-level features, thus inferring the continuous taillight state into: off, brake, left, right. Datasets including daytime, nighttime, congested road, highway, etc. were collected, tested and published in our work to demonstrate its effectiveness and practicability.
引用
收藏
页码:18088 / 18099
页数:12
相关论文
共 37 条
  • [1] Al-Maliki A., 2012, International Proceedings of Chemical, Biological and Environmental Engineering (IPCBEE), V37, P1
  • [2] Bi-LSTM-CRF Sequence Labeling for Keyphrase Extraction from Scholarly Documents
    Al-Zaidy, Rabah A.
    Caragea, Cornelia
    Giles, C. Lee
    [J]. WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 2551 - 2557
  • [3] Robust and Computationally Lightweight Autonomous Tracking of Vehicle Taillights and Signal Detection by Embedded Smart Cameras
    Almagambetov, Akhan
    Velipasalar, Senem
    Casares, Mauricio
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2015, 62 (06) : 3732 - 3741
  • [4] SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation
    Badrinarayanan, Vijay
    Kendall, Alex
    Cipolla, Roberto
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) : 2481 - 2495
  • [5] Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
  • [6] A Robust Algorithm for the Detection of Vehicle Turn Signals and Brake Lights
    Casares, Mauricio
    Almagambetov, Akhan
    Velipasalar, Senem
    [J]. 2012 IEEE NINTH INTERNATIONAL CONFERENCE ON ADVANCED VIDEO AND SIGNAL-BASED SURVEILLANCE (AVSS), 2012, : 386 - 391
  • [7] Driving style recognition method using braking characteristics based on hidden Markov model
    Deng, Chao
    Wu, Chaozhong
    Lyu, Nengchao
    Huang, Zhen
    [J]. PLOS ONE, 2017, 12 (08):
  • [8] Diba A., 2017, Temporal 3D ConvNets: New architecture and transfer learning for video classification
  • [9] Donahue J, 2015, PROC CVPR IEEE, P2625, DOI 10.1109/CVPR.2015.7298878
  • [10] Learning Spatiotemporal Features with 3D Convolutional Networks
    Du Tran
    Bourdev, Lubomir
    Fergus, Rob
    Torresani, Lorenzo
    Paluri, Manohar
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 4489 - 4497