Hallucination Reduction and Optimization for Large Language Model-Based Autonomous Driving

被引:0
|
作者
Wang, Jue [1 ]
机构
[1] Johns Hopkins Univ, Whiting Sch Engn, Baltimore, MD 21218 USA
来源
SYMMETRY-BASEL | 2024年 / 16卷 / 09期
关键词
autonomous driving; large language models; hallucination reduction;
D O I
10.3390/sym16091196
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Large language models (LLMs) are widely integrated into autonomous driving systems to enhance their operational intelligence and responsiveness and improve self-driving vehicles' overall performance. Despite these advances, LLMs still struggle between hallucinations-when models either misinterpret the environment or generate imaginary parts for downstream use cases-and taxing computational overhead that relegates their performance to strictly non-real-time operations. These are essential problems to solve to make autonomous driving as safe and efficient as possible. This work is thus focused on symmetrical trade-offs between the reduction of hallucination and optimization, leading to a framework for these two combined and at least specifically motivated by these limitations. This framework intends to generate a symmetry of mapping between real and virtual worlds. It helps in minimizing hallucinations and optimizing computational resource consumption reasonably. In autonomous driving tasks, we use multimodal LLMs that combine an image-encoding Visual Transformer (ViT) and a decoding GPT-2 with responses generated by the powerful new sequence generator from OpenAI known as GPT4. Our hallucination reduction and optimization framework leverages iterative refinement loops, RLHF-reinforcement learning from human feedback (RLHF)-along with symmetric performance metrics, e.g., BLEU, ROUGE, and CIDEr similarity scores between machine-generated answers specific to other human reference answers. This ensures that improvements in model accuracy are not overused to the detriment of increased computational overhead. Experimental results show a twofold improvement in decision-maker error rate and processing efficiency, resulting in an overall decrease of 30% for the model and a 25% improvement in processing efficiency across diverse driving scenarios. Not only does this symmetrical approach reduce hallucination, but it also better aligns the virtual and real-world representations.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] Model based vehicle detection and tracking for autonomous urban driving
    Petrovskaya, Anna
    Thrun, Sebastian
    AUTONOMOUS ROBOTS, 2009, 26 (2-3) : 123 - 139
  • [32] VWP:An Efficient DRL-Based Autonomous Driving Model
    Jin, Yan-Liang
    Ji, Ze-Yu
    Zeng, Dan
    Zhang, Xiao-Ping
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 2096 - 2108
  • [33] GPT-4 enhanced multimodal grounding for autonomous driving: Leveraging cross-modal attention with large language models
    Liao, Haicheng
    Shen, Huanming
    Li, Zhenning
    Wang, Chengyue
    Li, Guofa
    Bie, Yiming
    Xu, Chengzhong
    COMMUNICATIONS IN TRANSPORTATION RESEARCH, 2024, 4
  • [34] Towards Understanding Contracts Grammar: A Large Language Model-based Extractive Question-Answering Approach
    Rejithkumar, Gokul
    Anish, Preethu Rose
    Ghaisas, Smita
    32ND IEEE INTERNATIONAL REQUIREMENTS ENGINEERING CONFERENCE, RE 2024, 2024, : 310 - 320
  • [35] LAMARS: Large Language Model-Based Anticipation Mechanism Acceleration in Real-Time Robotic Systems
    Gao, Yifang
    Luo, Wei
    Wang, Xuye
    Zhang, Shunshun
    Goh, Patrick
    IEEE ACCESS, 2025, 13 : 3864 - 3880
  • [36] EPlus-LLM: A large language model-based computing platform for automated building energy modeling
    Jiang, Gang
    Ma, Zhihao
    Zhang, Liang
    Chen, Jianli
    APPLIED ENERGY, 2024, 367
  • [37] Large Language Model-Based Tuning Assistant for Variable Speed PMSM Drive with Cascade Control Structure
    Tarczewski, Tomasz
    Stojic, Djordje
    Dzielinski, Andrzej
    ELECTRONICS, 2025, 14 (02):
  • [38] From Playing the Story to Gaming the System: Repeat Experiences of a Large Language Model-Based Interactive Story
    Yong, Qing Ru
    Mitchell, Alex
    INTERACTIVE STORYTELLING, ICIDS 2023, PT I, 2023, 14383 : 395 - 409
  • [39] Exploring the potential of large language model-based chatbots in challenges of ribosome profiling data analysis: a review
    Ding, Zheyu
    Wei, Rong
    Xia, Jianing
    Mu, Yonghao
    Wang, Jiahuan
    Lin, Yingying
    BRIEFINGS IN BIOINFORMATICS, 2024, 26 (01)
  • [40] A Hybrid Machine Learning-based Control Strategy for Autonomous Driving Optimization
    Reda, Ahmad
    Benotsman, Rabab
    Bouzid, Ahmed
    Vasarhelyi, Jozsef
    ACTA POLYTECHNICA HUNGARICA, 2023, 20 (09) : 165 - 186