Chinese Generation and Security Index Evaluation Based on Large Language Model

被引:0
作者
Zhang, Yu [1 ]
Gao, Yongbing [1 ]
Li, Weihao [1 ]
Su, Zirong [1 ]
Yang, Lidong [1 ]
机构
[1] Inner Mongolia Univ Sci & Technol, Sch Numer Ind, Baotou, Inner Mongolia, Peoples R China
来源
2024 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING, IALP 2024 | 2024年
基金
中国国家自然科学基金;
关键词
Safety Assessment; Chinese Generation; AI Hallucination; Automatic Scoring; Large Language Model;
D O I
10.1109/IALP63756.2024.10661189
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study investigates the performance and security indicators of mainstream large language models in Chinese generation tasks. It explores potential security risks associated with these models and offers suggestions for improvement. The study utilizes publicly available datasets to assess Chinese language generation tasks, develops datasets and multidimensional security rating standards for security task evaluations, compares the performance of three models across 5 Chinese tasks and 6 security tasks, and conducts Pearson correlation analysis using GPT-4 and questionnaire surveys. Furthermore, the study implements automatic scoring based on GPT-3.5-Turbe. The experimental findings indicate that the models excel in Chinese language generation tasks. ERNIE Bot outperforms in the evaluation of ideology and ethics, ChatGPT excels in rumor and falsehood and privacy security assessments, and Claude performs well in assessing factual fallacy and social prejudice. The fine-tuned model demonstrates high accuracy in security tasks, yet all models exhibit security vulnerabilities. Integration into the prompt project proves to be effective in mitigating security risks. It is recommended that both domestic and foreign models adhere to the legal frameworks of each country, reduce AI hallucinations, continuously expand corpora, and update iterations accordingly.
引用
收藏
页码:151 / 161
页数:11
相关论文
共 36 条
  • [1] [Anonymous], 2021, C M ASS COMP LING AS
  • [2] Bai Y., 2022, arXiv, DOI DOI 10.48550/ARXIV.2204.05862
  • [3] Bai YT, 2022, Arxiv, DOI arXiv:2212.08073
  • [4] Bang Y, 2023, Arxiv, DOI [arXiv:2302.04023, DOI 10.48550/ARXIV.2302.04023]
  • [5] Bao SQ, 2020, Arxiv, DOI arXiv:1910.07931
  • [6] Brown TB, 2020, ADV NEUR IN, V33
  • [7] Chen S., 2023, J. Young Journalist, P57
  • [8] Das BC, 2024, Arxiv, DOI arXiv:2402.00888
  • [9] He W, 2018, Arxiv, DOI arXiv:1711.05073
  • [10] Hu B., 2015, J. Computer Science, P2667