Interpretable neural network classification model using first-order logic rules

IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Neurocomputing Pub Date : 2024-11-08 DOI:10.1016/j.neucom.2024.128840
Haiming Tuo, Zuqiang Meng, Zihao Shi, Daosheng Zhang
{"title":"Interpretable neural network classification model using first-order logic rules","authors":"Haiming Tuo,&nbsp;Zuqiang Meng,&nbsp;Zihao Shi,&nbsp;Daosheng Zhang","doi":"10.1016/j.neucom.2024.128840","DOIUrl":null,"url":null,"abstract":"<div><div>Over the past decade, the field of neural networks has made significant strides, particularly in deep learning. However, their limited interpretability has constrained their application in certain critical domains, drawing widespread criticism. Researchers have proposed various methods for explaining neural networks to address this challenge. This paper focuses on rule-based explanations for neural network classification problems. We propose IRCnet, a scalable classification model based on first-order logic rules. IRCnet consists of layers for learning conjunction and disjunction rules, utilizing binary logic activation functions to enhance interpretability. The model is initially trained using a continuous-weight version, which is later binarized to produce a discrete-weight version. During training, we innovatively employed gradient approximation method to handle the non-differentiable weight binarization function, thereby enabling the training of split matrices used for binarization. Finally, Conjunctive Normal Form (CNF) or Disjunctive Normal Form (DNF) rules are extracted from the model’s discrete-weight version. Experimental results indicate that our model achieves the highest or near-highest performance across various classification metrics in multiple structured datasets while demonstrating significant scalability. It effectively balances classification accuracy with the complexity of the generated rules.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128840"},"PeriodicalIF":5.5000,"publicationDate":"2024-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231224016114","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Over the past decade, the field of neural networks has made significant strides, particularly in deep learning. However, their limited interpretability has constrained their application in certain critical domains, drawing widespread criticism. Researchers have proposed various methods for explaining neural networks to address this challenge. This paper focuses on rule-based explanations for neural network classification problems. We propose IRCnet, a scalable classification model based on first-order logic rules. IRCnet consists of layers for learning conjunction and disjunction rules, utilizing binary logic activation functions to enhance interpretability. The model is initially trained using a continuous-weight version, which is later binarized to produce a discrete-weight version. During training, we innovatively employed gradient approximation method to handle the non-differentiable weight binarization function, thereby enabling the training of split matrices used for binarization. Finally, Conjunctive Normal Form (CNF) or Disjunctive Normal Form (DNF) rules are extracted from the model’s discrete-weight version. Experimental results indicate that our model achieves the highest or near-highest performance across various classification metrics in multiple structured datasets while demonstrating significant scalability. It effectively balances classification accuracy with the complexity of the generated rules.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
使用一阶逻辑规则的可解释神经网络分类模型
过去十年间,神经网络领域取得了长足的进步,尤其是在深度学习方面。然而,神经网络有限的可解释性限制了其在某些关键领域的应用,招致了广泛的批评。为了应对这一挑战,研究人员提出了各种解释神经网络的方法。本文重点关注神经网络分类问题的基于规则的解释。我们提出了基于一阶逻辑规则的可扩展分类模型 IRCnet。IRCnet 由用于学习连接和析取规则的层组成,利用二元逻辑激活函数来增强可解释性。该模型最初使用连续权重版本进行训练,然后将其二值化,生成离散权重版本。在训练过程中,我们创新性地采用了梯度逼近法来处理无差别权重二值化函数,从而实现了用于二值化的分割矩阵的训练。最后,从模型的离散权重版本中提取出连接正则表达式(CNF)或分离正则表达式(DNF)规则。实验结果表明,在多个结构化数据集中,我们的模型在各种分类指标上都取得了最高或接近最高的性能,同时还表现出显著的可扩展性。它有效地平衡了分类准确性和生成规则的复杂性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
期刊最新文献
Editorial Board Virtual sample generation for small sample learning: A survey, recent developments and future prospects Adaptive selection of spectral–spatial features for hyperspectral image classification using a modified-CBAM-based network FPGA-based component-wise LSTM training accelerator for neural granger causality analysis Multi-sensor information fusion in Internet of Vehicles based on deep learning: A review
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1