Exploring the capabilities of large language models for the generation of safety cases: the case of GPT-4

被引:1
作者
Sivakumar, Mithila [1 ]
Belle, Alvine Boaye [1 ]
Shan, Jinjun [1 ]
Shahandashti, Kimya Khakzad [1 ]
机构
[1] York Univ, Lassonde Sch Engn, Toronto, ON, Canada
来源
32ND INTERNATIONAL REQUIREMENTS ENGINEERING CONFERENCE WORKSHOPS, REW 2024 | 2024年
关键词
Safety cases; Safety assurance; Safety requirements; Machine learning; Large language models; Generative AI; RE quality models and their automation;
D O I
10.1109/REW61692.2024.00010
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The emergence of large language models (LLMs) and conversational interfaces, exemplified by ChatGPT, is nothing short of revolutionary. While their potential is undeniable across various domains, in this paper we investigate their uncharted territory-the exploration of generating safety cases. Our primary objective is to delve into the existing knowledge base of ChatGPT (specifically, GPT-4), focusing specifically on its understanding of the Goal Structuring Notation (GSN), a well-established notation allowing to visually represent safety cases. Subsequently, in our preliminary evaluation, we perform four distinct experiments with GPT-4. These experiments are designed to assess its capacity for generating safety cases within a defined system and application domain. To measure the performance of GPT-4 in this context, we compare the results it generates with ground-truth safety cases respectively created for an Xray system and a Machine Learning (ML)-enabled component for tire noise recognition (TNR) component in a vehicle. Our findings indicate that GPT-4 demonstrates the capacity to automatically produce safety arguments that are moderately accurate and reasonable. Furthermore, it also exhibits the capability to automatically generate safety cases whose semantic closely aligns with the one of the reference safety cases used as ground-truths in our experiments.
引用
收藏
页码:35 / 45
页数:11
相关论文
共 35 条
[1]   Using Assurance Cases to Prevent Malicious Behaviour from Targeting Safety Vulnerabilities [J].
Bandur, Victor ;
Lawford, Mark ;
Mosser, Sebastien ;
Paige, Richard F. ;
Pantelic, Vera ;
Wassyng, Alan .
COMPUTER SAFETY, RELIABILITY, AND SECURITY, SAFECOMP 2023 WORKSHOPS, 2023, 14182 :5-14
[2]   Position paper: a vision for the dynamic safety assurance of ML-enabled autonomous driving systems [J].
Belle, Alvine Boaye ;
Hemmati, Hadi ;
Lethbridge, Timothy C. .
2023 IEEE 31ST INTERNATIONAL REQUIREMENTS ENGINEERING CONFERENCE WORKSHOPS, REW, 2023, :297-301
[3]   Safety Assurance of Machine Learning for Chassis Control Functions [J].
Burton, Simon ;
Kurzidem, Iwo ;
Schwaiger, Adrian ;
Schleiss, Philipp ;
Unterreiner, Michael ;
Graeber, Torben ;
Becker, Philipp .
COMPUTER SAFETY, RELIABILITY, AND SECURITY (SAFECOMP 2021), 2021, 12852 :149-162
[4]   Graphical safety assurance case using Goal Structuring Notation (GSN)-challenges, opportunities and a framework for autonomous trains [J].
Chelouati, Mohammed ;
Boussif, Abderraouf ;
Beugin, Julie ;
El Koursi, El-Miloudi .
RELIABILITY ENGINEERING & SYSTEM SAFETY, 2023, 230
[5]   On the Use of GPT-4 for Creating Goal Models: An Exploratory Study [J].
Chen, Boqi ;
Chen, Kua ;
Hassani, Shabnam ;
Yang, Yujing ;
Amyot, Daniel ;
Lessard, Lysanne ;
Mussbachcr, Gunter ;
Sabetzadeh, Mehrdad ;
Varro, Daniel .
2023 IEEE 31ST INTERNATIONAL REQUIREMENTS ENGINEERING CONFERENCE WORKSHOPS, REW, 2023, :262-271
[6]   Model-based safety engineering for autonomous train map [J].
Chouchani, Nadia ;
Debbech, Sana ;
Perin, Matthieu .
JOURNAL OF SYSTEMS AND SOFTWARE, 2022, 183
[7]   Automating the Assembly of Aviation Safety Cases [J].
Denney, Ewen ;
Pai, Ganesh .
IEEE TRANSACTIONS ON RELIABILITY, 2014, 63 (04) :830-849
[8]   Representation of Confidence in Assurance Cases using the Beta Distribution [J].
Duan, Lian ;
Rayadurgam, Sanjai ;
Heimdahl, Mats ;
Sokolsky, Oleg ;
Lee, Insup .
2016 IEEE 17TH INTERNATIONAL SYMPOSIUM ON HIGH ASSURANCE SYSTEMS ENGINEERING (HASE), 2016, :86-93
[9]  
Holloway C.M., 2008, 2008 3 IET INT C SYS, P1
[10]   ChatGPT for good? On opportunities and challenges of large language models for education [J].
Kasneci, Enkelejda ;
Sessler, Kathrin ;
Kuechemann, Stefan ;
Bannert, Maria ;
Dementieva, Daryna ;
Fischer, Frank ;
Gasser, Urs ;
Groh, Georg ;
Guennemann, Stephan ;
Huellermeier, Eyke ;
Krusche, Stepha ;
Kutyniok, Gitta ;
Michaeli, Tilman ;
Nerdel, Claudia ;
Pfeffer, Juergen ;
Poquet, Oleksandra ;
Sailer, Michael ;
Schmidt, Albrecht ;
Seidel, Tina ;
Stadler, Matthias ;
Weller, Jochen ;
Kuhn, Jochen ;
Kasneci, Gjergji .
LEARNING AND INDIVIDUAL DIFFERENCES, 2023, 103