A Survey on Verification and Validation, Testing and Evaluations of Neurosymbolic Artificial Intelligence

被引:4
作者
Renkhoff J. [1 ]
Feng K. [2 ]
Meier-Doernberg M. [2 ]
Velasquez A. [3 ]
Song H.H. [1 ]
机构
[1] University of Maryland, Security and Optimization for Networked Globe Laboratory (SONG Lab), Department of Information Systems, Baltimore, 21250, MD
[2] Embry-Riddle Aeronautical University, Department of Electrical Engineering and Computer Science, Daytona Beach, 32114, FL
[3] University of Colorado, Department of Computer Science, Boulder, 80309, CO
来源
IEEE Transactions on Artificial Intelligence | 2024年 / 5卷 / 08期
基金
美国国家科学基金会;
关键词
Deep learning (DL); evaluation; neurosymbolic artificial intelligence (AI); safety; security; testing; trustworthiness; validation; verification;
D O I
10.1109/TAI.2024.3351798
中图分类号
学科分类号
摘要
Neurosymbolic artificial intelligence (AI) is an emerging branch of AI that combines the strengths of symbolic AI and subsymbolic AI. Symbolic AI is based on the idea that intelligence can be represented using semantically meaningful symbolic rules and representations, while deep learning (DL), or sometimes called subsymbolic AI, is based on the idea that intelligence emerges from the collective behavior of artificial neurons that are connected to each other. A major drawback of DL is that it acts as a 'black box,' meaning that predictions are difficult to explain, making the testing & evaluation (T&E) and validation & verification (V&V) processes of a system that uses subsymbolic AI a challenge. Since neurosymbolic AI combines the advantages of both symbolic and subsymbolic AI, this survey explores how neurosymbolic applications can ease the V&V process. This survey considers two taxonomies of neurosymbolic AI, evaluates them, and analyzes which algorithms are commonly used as the symbolic and subsymbolic components in current applications. Additionally, an overview of current techniques for the T&E and V&V processes of these components is provided. Furthermore, it is investigated how the symbolic part is used for T&E and V&V purposes in current neurosymbolic applications. Our research shows that neurosymbolic AI has great potential to ease the T&E and V&V processes of subsymbolic AI by leveraging the possibilities of symbolic AI. Additionally, the applicability of current T&E and V&V methods to neurosymbolic AI is assessed, and how different neurosymbolic architectures can impact these methods is explored. It is found that current T&E and V&V techniques are partly sufficient to test, evaluate, verify, or validate the symbolic and subsymbolic part of neurosymbolic applications independently, while some of them use approaches where current T&E and V&V methods are not applicable by default, and adjustments or even new approaches are needed. Our research shows that there is great potential in using symbolic AI to test, evaluate, verify, or validate the predictions of a subsymbolic model, making neurosymbolic AI an interesting research direction for safe, secure, and trustworthy AI. © 2020 IEEE.
引用
收藏
页码:3765 / 3779
页数:14
相关论文
共 125 条
  • [1] Garcez A.D., Lamb L.C., Neurosymbolic AI: The 3rd wave, Artif. Intell. Rev., 56, pp. 1-20, (2023)
  • [2] Bishop C.M., Neural networks and their applications, Rev. Sci. Instrum., 65, 6, pp. 1803-1832, (1994)
  • [3] Garnelo M., Shanahan M., Reconciling deep learning with symbolic artificial intelligence: Representing objects and relations, Curr. Opin. Behav. Sci., 29, pp. 17-23, (2019)
  • [4] Acharya K., Raza W., Dourado C., Velasquez A., Song H.H., Neurosymbolic reinforcement learning and planning: A survey, IEEE Trans. Artif. Intell., early access, (2023)
  • [5] Velasquez A., Transfer from imprecise and abstract models to autonomous technologies (TIAMAT), (2023)
  • [6] Biden R., Executive order on the safe, secure, and trustworthy development and use of artificial intelligence, (2023)
  • [7] Huang X., Safety and reliability of deep learning: (Brief overview), Proc. 1st Int. Workshop Verification Auton. Rob. Syst. (VARS)., (2021)
  • [8] Huang X., Et al., A survey of safety and trustworthiness of deep neural networks: Verification, testing, adversarial attack and defence, and interpretability, Comput. Sci. Rev., 37, (2020)
  • [9] Wallace D., Fujii R., Software verification and validation: Its role in computer assurance and its relationship with software project management standards, (2024)
  • [10] Sargent R.G., A tutorial on validation and verification of simulation models, Proc. 20th Conf. Winter Simul., pp. 33-39, (1988)