Vector-Based Semantic Scenario Search for Vehicular Traffic

被引:0
作者
Bhoomika, A. P. [1 ]
Srinivasa, Srinath [1 ]
Indla, Vijaya Sarathi [2 ]
Mukherjee, Saikat [2 ]
机构
[1] Int Inst Informat Technol, Bengaluru, India
[2] Siemens Technol & Serv Private Ltd, Bengaluru, India
来源
BIG DATA ANALYTICS IN ASTRONOMY, SCIENCE, AND ENGINEERING, BDA 2023 | 2024年 / 14516卷
关键词
Autonomous Vehicles; Image Captioning; Vector Embeddings; Semantic Scenario search;
D O I
10.1007/978-3-031-58502-9_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autonomous Vehicles (AVs) are expected to have the potential to impact urban mobility by providing increased safety, reducing traffic congestion, mitigating accidents and reducing emissions. Since AVs operate with little or no human intervention, it is very essential to perceive the external world and understand different objects and their relationships in the scene, and respond appropriately. For doing this effectively, AVs need to be trained on a variety of traffic situations and appropriate responses to them. Behaviour of vehicular traffic varies widely from one part of the world to another. An AV trained for traffic conditions in one part of the world may not be effective, or worse, even be risky in some other part of the world. There is hence a need to create datasets of vehicular traffic scenarios and design mechanisms to query, retrieve and reason about dynamic traffic scenarios. This paper discusses a method for vector based scenario search using a natural language interface for describing traffic scenarios. We first generate textual descriptions of snapshots of traffic scenarios captured from instrumenting vehicles using image captioning libraries. Next, we create vector embeddings of the captions, store them in a vector database to enable semantic scenario search using natural language based queries. This is an ongoing work where different other modalities of scenarios data are planned to be supported over an underlying image captioning and natural language search interface. Experimental results on the image captioning core, show encouraging results.
引用
收藏
页码:160 / 171
页数:12
相关论文
共 21 条
[1]   A Survey of Autonomous Vehicles: Enabling Communication Technologies and Challenges [J].
Ahangar, M. Nadeem ;
Ahmed, Qasim Z. ;
Khan, Fahd A. ;
Hafeez, Maryam .
SENSORS, 2021, 21 (03) :1-33
[2]   nuScenes: A multimodal dataset for autonomous driving [J].
Caesar, Holger ;
Bankiti, Varun ;
Lang, Alex H. ;
Vora, Sourabh ;
Liong, Venice Erin ;
Xu, Qiang ;
Krishnan, Anush ;
Pan, Yu ;
Baldan, Giancarlo ;
Beijbom, Oscar .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :11618-11628
[3]  
Denkowski Michael., 2014, P 9 WORKSHOP STAT MA, P376
[4]  
Han YK, 2023, Arxiv, DOI arXiv:2310.11703
[5]   A Comprehensive Survey of Deep Learning for Image Captioning [J].
Hossain, Md Zakir ;
Sohel, Ferdous ;
Shiratuddin, Mohd Fairuz ;
Laga, Hamid .
ACM COMPUTING SURVEYS, 2019, 51 (06)
[6]  
Li JN, 2022, PR MACH LEARN RES
[7]  
Li JN, 2023, Arxiv, DOI arXiv:2301.12597
[8]   The Traffic Scene Understanding and Prediction Based on Image Captioning [J].
Li, Wei ;
Qu, Zhaowei ;
Song, Haiyu ;
Wang, Pengjie ;
Xue, Bo .
IEEE ACCESS, 2021, 9 :1420-1427
[9]   Caption Generation From Road Images for Traffic Scene Modeling [J].
Li, Yaochen ;
Wu, Chuan ;
Li, Ling ;
Liu, Yuehu ;
Zhu, Jihua .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (07) :7805-7816
[10]  
Lin Chin-Yew., 2004, P ACL WORKSHOP TEXT, P74, DOI DOI 10.1179/CIM.2004.5