LimSim plus plus : A Closed-Loop Platform for Deploying Multimodal LLMs in Autonomous Driving

被引:2
作者
Fu, Daocheng [1 ]
Lei, Wenjie [1 ,2 ]
Wen, Licheng [1 ]
Cai, Pinlong [1 ]
Mao, Song [1 ]
Dou, Min [1 ]
Shi, Botian [1 ]
Qiao, Yu [1 ]
机构
[1] Shanghai Artificial Intelligence Lab, Shanghai, Peoples R China
[2] Zhejiang Univ, Coll Control Sci & Engn, Hangzhou, Peoples R China
来源
2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024 | 2024年
基金
国家重点研发计划;
关键词
D O I
10.1109/IV55156.2024.10588848
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The emergence of Multimodal Large Language Models ((M)LLMs) has ushered in new avenues in artificial intelligence, particularly for autonomous driving by offering enhanced understanding and reasoning capabilities. This paper introduces LimSim++, an extended version of LimSim designed for the application of (M)LLMs in autonomous driving. Acknowledging the limitations of existing simulation platforms, LimSim++ addresses the need for a long-term closed-loop infrastructure supporting continuous learning and improved generalization in autonomous driving. The platform offers extended-duration, multi-scenario simulations, providing crucial information for (M)LLM-driven vehicles. Users can engage in prompt engineering, model evaluation, and framework enhancement, making LimSim++ a versatile tool for research and practice. This paper additionally introduces a baseline (M)LLM-driven framework, systematically validated through quantitative experiments across diverse scenarios. The open-source resources of LimSim++ are available at: https://pjlabadg.github.io/limsim-plus/.
引用
收藏
页码:1084 / 1090
页数:7
相关论文
共 38 条
[1]  
Bae I., 2020, ARXIV
[2]  
Caesar H., 2021, arXiv
[3]   nuScenes: A multimodal dataset for autonomous driving [J].
Caesar, Holger ;
Bankiti, Varun ;
Lang, Alex H. ;
Vora, Sourabh ;
Liong, Venice Erin ;
Xu, Qiang ;
Krishnan, Anush ;
Pan, Yu ;
Baldan, Giancarlo ;
Beijbom, Oscar .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :11618-11628
[4]  
Chen L., 2023, ARXIV
[5]  
Chen Lin, 2023, arXiv
[6]  
Chowdhery A, 2023, J MACH LEARN RES, V24
[7]  
Cui C, 2023, ARXIV
[8]   Drive as You Speak: Enabling Human-Like Interaction with Large Language Models in Autonomous Vehicles [J].
Cui, Can ;
Ma, Yunsheng ;
Cao, Xu ;
Ye, Wenqian ;
Wang, Ziran .
2024 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS, WACVW 2024, 2024, :902-909
[9]   A Survey on Multimodal Large Language Models for Autonomous Driving [J].
Cui, Can ;
Ma, Yunsheng ;
Cao, Xu ;
Ye, Wenqian ;
Zhou, Yang ;
Liang, Kaizhao ;
Chen, Jintai ;
Lu, Juanwu ;
Yang, Zichong ;
Liao, Kuei-Da ;
Gao, Tianren ;
Li, Erlong ;
Tang, Kun ;
Cao, Zhipeng ;
Zhou, Tong ;
Liu, Ao ;
Yan, Xinrui ;
Mei, Shuqi ;
Cao, Jianguo ;
Wang, Ziran ;
Zheng, Chao .
2024 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS, WACVW 2024, 2024, :958-979
[10]  
Dosovitskiy A, 2017, PR MACH LEARN RES, V78