Interpretable neural network classification model using first-order logic rules☆

被引:0
作者
Tuo, Haiming [1 ]
Meng, Zuqiang [1 ]
Shi, Zihao [1 ]
Zhang, Daosheng [1 ]
机构
[1] Guangxi Univ, Sch Comp & Elect & Informat, Nanning 530004, Peoples R China
基金
中国国家自然科学基金;
关键词
Interpretable; FOL rules; Gradient approximation; Classification;
D O I
10.1016/j.neucom.2024.128840
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Over the past decade, the field of neural networks has made significant strides, particularly in deep learning. However, their limited interpretability has constrained their application in certain critical domains, drawing widespread criticism. Researchers have proposed various methods for explaining neural networks to address this challenge. This paper focuses on rule-based explanations for neural network classification problems. We propose IRCnet, a scalable classification model based on first-order logic rules. IRCnet consists of layers for learning conjunction and disjunction rules, utilizing binary logic activation functions to enhance interpretability. The model is initially trained using a continuous-weight version, which is later binarized to produce a discrete-weight version. During training, we innovatively employed gradient approximation method to handle the non-differentiable weight binarization function, thereby enabling the training of split matrices used for binarization. Finally, Conjunctive Normal Form (CNF) or Disjunctive Normal Form (DNF) rules are extracted from the model's discrete-weight version. Experimental results indicate that our model achieves the highest or near-highest performance across various classification metrics in multiple structured datasets while demonstrating significant scalability. It effectively balances classification accuracy with the complexity of the generated rules.
引用
收藏
页数:18
相关论文
共 51 条
  • [1] Survey and critique of techniques for extracting rules from trained artificial neural networks
    Andrews, R
    Diederich, J
    Tickle, AB
    [J]. KNOWLEDGE-BASED SYSTEMS, 1995, 8 (06) : 373 - 389
  • [2] [Anonymous], 1952, Introduction to Metamathematics
  • [3] Arya V, 2019, Arxiv, DOI arXiv:1909.03012
  • [4] Bacon A., 2023, A Philosophical Introduction to Higher-Order Logics, DOI DOI 10.4324/9781003039181
  • [5] Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
    Barredo Arrieta, Alejandro
    Diaz-Rodriguez, Natalia
    Del Ser, Javier
    Bennetot, Adrien
    Tabik, Siham
    Barbado, Alberto
    Garcia, Salvador
    Gil-Lopez, Sergio
    Molina, Daniel
    Benjamins, Richard
    Chatila, Raja
    Herrera, Francisco
    [J]. INFORMATION FUSION, 2020, 58 : 82 - 115
  • [6] Optimal classification trees
    Bertsimas, Dimitris
    Dunn, Jack
    [J]. MACHINE LEARNING, 2017, 106 (07) : 1039 - 1082
  • [7] Breiman L., 2017, Classification and Regression Trees, DOI DOI 10.1201/9781315139470
  • [8] Chen CF, 2019, ADV NEUR IN, V32
  • [9] Training threshold neural networks by extreme learning machine and adaptive stochastic resonance
    Chen, Zejia
    Duan, Fabing
    Chapeau-Blondeau, Francois
    Abbott, Derek
    [J]. PHYSICS LETTERS A, 2022, 432
  • [10] Logic Explained Networks
    Ciravegna, Gabriele
    Barbiero, Pietro
    Giannini, Francesco
    Gori, Marco
    Lio, Pietro
    Maggini, Marco
    Melacci, Stefano
    [J]. ARTIFICIAL INTELLIGENCE, 2023, 314