Hallucination Reduction and Optimization for Large Language Model-Based Autonomous Driving

被引:0
|
作者
Wang, Jue [1 ]
机构
[1] Johns Hopkins Univ, Whiting Sch Engn, Baltimore, MD 21218 USA
来源
SYMMETRY-BASEL | 2024年 / 16卷 / 09期
关键词
autonomous driving; large language models; hallucination reduction;
D O I
10.3390/sym16091196
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Large language models (LLMs) are widely integrated into autonomous driving systems to enhance their operational intelligence and responsiveness and improve self-driving vehicles' overall performance. Despite these advances, LLMs still struggle between hallucinations-when models either misinterpret the environment or generate imaginary parts for downstream use cases-and taxing computational overhead that relegates their performance to strictly non-real-time operations. These are essential problems to solve to make autonomous driving as safe and efficient as possible. This work is thus focused on symmetrical trade-offs between the reduction of hallucination and optimization, leading to a framework for these two combined and at least specifically motivated by these limitations. This framework intends to generate a symmetry of mapping between real and virtual worlds. It helps in minimizing hallucinations and optimizing computational resource consumption reasonably. In autonomous driving tasks, we use multimodal LLMs that combine an image-encoding Visual Transformer (ViT) and a decoding GPT-2 with responses generated by the powerful new sequence generator from OpenAI known as GPT4. Our hallucination reduction and optimization framework leverages iterative refinement loops, RLHF-reinforcement learning from human feedback (RLHF)-along with symmetric performance metrics, e.g., BLEU, ROUGE, and CIDEr similarity scores between machine-generated answers specific to other human reference answers. This ensures that improvements in model accuracy are not overused to the detriment of increased computational overhead. Experimental results show a twofold improvement in decision-maker error rate and processing efficiency, resulting in an overall decrease of 30% for the model and a 25% improvement in processing efficiency across diverse driving scenarios. Not only does this symmetrical approach reduce hallucination, but it also better aligns the virtual and real-world representations.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Asynchronous Large Language Model Enhanced Planner for Autonomous Driving
    Chen, Yuan
    Ding, Zi-han
    Wang, Ziqin
    Wang, Yan
    Zhang, Lijun
    Liu, Si
    COMPUTER VISION - ECCV 2024, PT XXXVI, 2025, 15094 : 22 - 38
  • [2] Restricted Natural Language and Model-based Adaptive Test Generation for Autonomous Driving
    Shi, Yize
    Lu, Chengjie
    Zhang, Man
    Zhang, Huihui
    Yue, Tao
    Ali, Shaukat
    24TH INTERNATIONAL CONFERENCE ON MODEL-DRIVEN ENGINEERING LANGUAGES AND SYSTEMS (MODELS 2021), 2021, : 101 - 111
  • [3] Bidirectional Planning for Autonomous Driving Framework with Large Language Model
    Ma, Zhikun
    Sun, Qicong
    Matsumaru, Takafumi
    SENSORS, 2024, 24 (20)
  • [4] Facilitating Autonomous Driving Tasks With Large Language Models
    Wu, Mengyao
    Yu, F. Richard
    Liu, Peter Xiaoping
    He, Ying
    IEEE INTELLIGENT SYSTEMS, 2025, 40 (01) : 45 - 52
  • [5] DriveLLM: Charting the Path Toward Full Autonomous Driving With Large Language Models
    Cui, Yaodong
    Huang, Shucheng
    Zhong, Jiaming
    Liu, Zhenan
    Wang, Yutong
    Sun, Chen
    Li, Bai
    Wang, Xiao
    Khajepour, Amir
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (01): : 1450 - 1464
  • [6] Large language model-based evolutionary optimizer: Reasoning with elitism
    Brahmachary, Shuvayan
    Joshi, Subodh M.
    Panda, Aniruddha
    Koneripalli, Kaushik
    Sagotra, Arun Kumar
    Patel, Harshil
    Sharma, Ankush
    Jagtap, Ameya D.
    Kalyanaraman, Kaushic
    NEUROCOMPUTING, 2025, 622
  • [7] Uncertainty-Aware Model-Based Reinforcement Learning: Methodology and Application in Autonomous Driving
    Wu, Jingda
    Huang, Zhiyu
    Lv, Chen
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (01): : 194 - 203
  • [8] Domain Knowledge Distillation from Large Language Model: An Empirical Study in the Autonomous Driving Domain
    Tang, Yun
    da Costa, Antonio A. Bruto
    Zhang, Xizhe
    Patrick, Irvine
    Khastgir, Siddartha
    Jennings, Paul
    2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC, 2023, : 3893 - 3900
  • [9] Large Language Model-based Test Case Generation for GP Agents
    Jorgensen, Steven
    Nadizar, Giorgia
    Pietropolli, Gloria
    Manzoni, Luca
    Medvet, Eric
    O'Reilly, Una-May
    Hemberg, Erik
    PROCEEDINGS OF THE 2024 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, GECCO 2024, 2024, : 914 - 923
  • [10] LUNA: A Model-Based Universal Analysis Framework for Large Language Models
    Song, Da
    Xie, Xuan
    Song, Jiayang
    Zhu, Derui
    Huang, Yuheng
    Felix, Juefei-Xu
    Ma, Lei
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2024, 50 (07) : 1921 - 1948