A Survey of Metrics to Enhance Training Dependability in Large Language Models

被引:0
|
作者
Fang, Wenyi [1 ]
Zhang, Hao [1 ]
Gong, Ziyu [1 ]
Zeng, Longbin [1 ]
Lu, Xuhui [1 ,2 ]
Liu, Biao [1 ]
Wu, Xiaoyu [1 ]
Zheng, Yang [1 ]
Hu, Zheng [1 ]
Zhang, Xun [1 ]
机构
[1] Huawei Technol Co Ltd, Shenzhen, Peoples R China
[2] Beihang Univ, Sch Automat Sci & Elect Engn, Beijing, Peoples R China
来源
2023 IEEE 34TH INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING WORKSHOPS, ISSREW | 2023年
关键词
Large Language Model; Dependability; Monitoring Metric;
D O I
10.1109/ISSREW60843.2023.00071
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The rapidly advancing field of artificial intelligence requires meticulous attention to the training and monitoring of large language models (LLMs). This paper offers a systematic analysis of existing metrics and introduces new ones, focusing on their theoretical underpinnings and practical implementations. We present empirical results and insights into the performance of selected metrics, elucidating the complex interplay of variables in the training process. Our comprehensive approach provides significant insights into LLM training, and promises to improve the dependability and efficiency of future models.
引用
收藏
页码:180 / 185
页数:6
相关论文
共 50 条
  • [1] A survey of multilingual large language models
    Qin, Libo
    Chen, Qiguang
    Zhou, Yuhang
    Chen, Zhi
    Li, Yinghui
    Liao, Lizi
    Li, Min
    Che, Wanxiang
    Yu, Philip S.
    PATTERNS, 2025, 6 (01):
  • [2] A survey on multimodal large language models
    Yin, Shukang
    Fu, Chaoyou
    Zhao, Sirui
    Li, Ke
    Sun, Xing
    Xu, Tong
    Chen, Enhong
    NATIONAL SCIENCE REVIEW, 2024, 11 (12)
  • [3] A comprehensive survey of large language models and multimodal large models in medicine
    Xiao, Hanguang
    Zhou, Feizhong
    Liu, Xingyue
    Liu, Tianqi
    Li, Zhipeng
    Liu, Xin
    Huang, Xiaoxuan
    INFORMATION FUSION, 2025, 117
  • [4] Targeted training for numerical reasoning with large language models
    Li, Xiao
    Liu, Sichen
    Zhu, Yin
    Cheng, Gong
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, 67 (01) : 197 - 221
  • [5] A survey analysis of the adoption of large language models among pathologists
    Laohawetwanit, Thiyaphat
    Pinto, Daniel Gomes
    Bychkov, Andrey
    AMERICAN JOURNAL OF CLINICAL PATHOLOGY, 2024, 163 (01) : 52 - 59
  • [6] History, development, and principles of large language models: an introductory survey
    Zichong Wang
    Zhibo Chu
    Thang Viet Doan
    Shiwen Ni
    Min Yang
    Wenbin Zhang
    AI and Ethics, 2025, 5 (3): : 1955 - 1971
  • [7] Dermatologists' Perspectives and Usage of Large Language Models in Practice: An Exploratory Survey
    Gui, Haiwen
    Rezaei, Shawheen J.
    Schlessinger, Daniel
    Weed, Jason
    Lester, Jenna
    Wongvibulsin, Shannon
    Mitchell, Dom
    Ko, Justin
    Rotemberg, Veronica
    Lee, Ivy
    Daneshjou, Roxana
    JOURNAL OF INVESTIGATIVE DERMATOLOGY, 2024, 144 (10) : 2298 - 2301
  • [8] A Survey: Collaborative Hardware and Software Design in the Era of Large Language Models
    Guo, Cong
    Cheng, Feng
    Du, Zhixu
    Kiessling, James
    Ku, Jonathan
    Li, Shiyu
    Li, Ziru
    Ma, Mingyuan
    Molom-Ochir, Tergel
    Morris, Benjamin
    Shan, Haoxuan
    Sun, Jingwei
    Wang, Yitu
    Wei, Chiyue
    Wu, Xueying
    Wu, Yuhao
    Yang, Hao Frank
    Zhang, Jingyang
    Zhang, Junyao
    Zheng, Qilin
    Zhou, Guanglei
    Li, Hai
    Chen, Yiran
    IEEE CIRCUITS AND SYSTEMS MAGAZINE, 2025, 25 (01) : 35 - 57
  • [9] Instruct Large Language Models to Generate Scientific Literature Survey Step by Step
    Lai, Yuxuan
    Wu, Yupeng
    Wang, Yidan
    Hu, Wenpeng
    Zheng, Chen
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT V, NLPCC 2024, 2025, 15363 : 484 - 496
  • [10] Parameter-efficient fine-tuning in large language models: a survey of methodologies
    Wang, Luping
    Chen, Sheng
    Jiang, Linnan
    Pan, Shu
    Cai, Runze
    Yang, Sen
    Yang, Fei
    ARTIFICIAL INTELLIGENCE REVIEW, 2025, 58 (08)