BEVSeg2GTA: Joint Vehicle Segmentation and Graph Neural Networks for Ego Vehicle Trajectory Prediction in Bird's-Eye-View

被引:0
|
作者
Sharma, Sushil [1 ,2 ,3 ]
Das, Arindam [1 ,3 ]
Sistu, Ganesh [1 ,3 ]
Halton, Mark [1 ,3 ]
Eising, Ciaran [1 ,2 ,3 ]
机构
[1] Univ Limerick, Dept Elect & Comp Engn, Limerick V94 T9PX, Ireland
[2] Univ Limerick, SFI CRT Fdn Data Sci, Limerick V94 T9PX, Ireland
[3] Univ Limerick, Data Driven Comp Engn D2iCE Res Ctr, Limerick V94 T9PX, Ireland
来源
IEEE ACCESS | 2024年 / 12卷
基金
爱尔兰科学基金会;
关键词
Trajectory; Cameras; Transformers; Autonomous vehicles; Probabilistic logic; Accuracy; Graph neural networks; Encoding; Spatial temporal resolution; Multi-view camera; encoder-decoder transformer; Bird's-Eye-View; semantic segmentation; graph neural network; spatio-temporal probabilistic network; trajectory prediction; SEMANTIC SEGMENTATION;
D O I
10.1109/ACCESS.2024.3459595
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Predicting the trajectory of the ego vehicle is a critical task for autonomous vehicles. Even though traffic regulations have defined boundaries, various behaviors of the agents in real-life situations introduce complexities that are hard to capture comprehensively. This has led to a rising curiosity in ego vehicle trajectory prediction based on learning techniques. In this paper, we introduce BEVSeg2GTA (Bird's-Eye-View Joint Vehicle Segmentation and Graph Neural Network Trajectory Prediction), a novel approach that aims to forecast trajectories by treating perception and trajectory prediction as interconnected elements of a single system. By integrating these tasks, we demonstrate the possibility of improving perception accuracy and trajectory prediction error. Initially, an encoder-decoder transformer-based deep network is employed to convert the multi-view camera images to a Bird's-Eye-View representation followed by semantic segmentation of crucial agents, including the ego vehicle, other vehicles, and pedestrians within the scene. Integrating state-of-the-art backbone (such as EfficientNet) facilitates the extraction of strong features, which are used to construct a graph wherein a node represents each object within the scene. Subsequently, the connections of these nodes are established by a k-Nearest Neighbors algorithm based on the distance metric. Further, the node and image features are fed into a Graph Neural Network to learn the complex relationships between agents in a spatial context. Finally, the Graph Neural Network learned features are passed to a Spatio-Temporal Probabilistic Network to predict the ego vehicle's future trajectory accurately. The proposed framework, BEVSeg2GTA, has been extensively evaluated on nuScenes datasets. The results demonstrate that the proposed method improves the state-of-the-art performance.
引用
收藏
页码:132159 / 132174
页数:16
相关论文
共 6 条
  • [1] BEVSeg: Geometry and Data-Driven Based Multi-view Segmentation in Bird's-Eye-View
    Chen, Qiuxiao
    Tai, Hung-Shuo
    Li, Pengfei
    Wang, Ke
    Qi, Xiaojun
    COMPUTER VISION SYSTEMS, ICVS 2023, 2023, 14253 : 432 - 443
  • [2] Vehicle trajectory prediction on highways using bird eye view representations and deep learning
    Izquierdo, Ruben
    Quintanar, Alvaro
    Fernandez Llorca, David
    Garcia Daza, Ivan
    Hernandez, Noelia
    Parra, Ignacio
    Angel Sotelo, Miguel
    APPLIED INTELLIGENCE, 2023, 53 (07) : 8370 - 8388
  • [3] Vehicle trajectory prediction on highways using bird eye view representations and deep learning
    Rubén Izquierdo
    Álvaro Quintanar
    David Fernández Llorca
    Iván García Daza
    Noelia Hernández
    Ignacio Parra
    Miguel Ángel Sotelo
    Applied Intelligence, 2023, 53 : 8370 - 8388
  • [4] Exploring the Impact of Social Interaction Factors on Vehicle Trajectory Prediction with Graph Neural Networks
    Gao, Tianyang
    Qiu, Chunlong
    Liu, Shuaipeng
    Zhuang, Weichao
    2024 10TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, CONTROL AND ROBOTICS, EECR 2024, 2024, : 208 - 214
  • [5] Graph-Based Interaction-Aware Multimodal 2D Vehicle Trajectory Prediction Using Diffusion Graph Convolutional Networks
    Wu, Keshu
    Zhou, Yang
    Shi, Haotian
    Li, Xiaopeng
    Ran, Bin
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (02): : 3630 - 3643
  • [6] Time-sensitive and Distance-tolerant Deep Learning-Based Vehicle Detection Using High-Resolution Radar Bird's-Eye-View Images
    Zheng, Ruxin
    Sun, Shunqiao
    Liu, Hongshan
    Wu, Teresa
    2023 IEEE RADAR CONFERENCE, RADARCONF23, 2023,