Yuqi Wang, Dawei Dai, Da Liu, Shuyin Xia, Guoyin Wang
{"title":"BTSC: Binary tree structure convolution layers for building interpretable decision-making deep CNN","authors":"Yuqi Wang, Dawei Dai, Da Liu, Shuyin Xia, Guoyin Wang","doi":"10.1049/cit2.12328","DOIUrl":null,"url":null,"abstract":"<p>Although deep convolution neural network (DCNN) has achieved great success in computer vision field, such models are considered to lack interpretability in decision-making. One of fundamental issues is that its decision mechanism is considered to be a “black-box” operation. The authors design the binary tree structure convolution (BTSC) module and control the activation level of particular neurons to build the interpretable DCNN model. First, the authors design a BTSC module, in which each parent node generates two independent child layers, and then integrate them into a normal DCNN model. The main advantages of the BTSC are as follows: 1) child nodes of the different parent nodes do not interfere with each other; 2) parent and child nodes can inherit knowledge. Second, considering the activation level of neurons, the authors design an information coding objective to guide neural nodes to learn the particular information coding that is expected. Through the experiments, the authors can verify that: 1) the decision-making made by both the ResNet and DenseNet models can be explained well based on the \"decision information flow path\" (known as <b>the decision-path</b>) formed in the BTSC module; 2) <b>the decision-path</b> can reasonably interpret the decision reversal mechanism (Robustness mechanism) of the DCNN model; 3) the credibility of decision-making can be measured by the matching degree between the actual and expected <b>decision-path</b>.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"9 5","pages":"1331-1345"},"PeriodicalIF":8.4000,"publicationDate":"2024-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12328","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"CAAI Transactions on Intelligence Technology","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1049/cit2.12328","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Although deep convolution neural network (DCNN) has achieved great success in computer vision field, such models are considered to lack interpretability in decision-making. One of fundamental issues is that its decision mechanism is considered to be a “black-box” operation. The authors design the binary tree structure convolution (BTSC) module and control the activation level of particular neurons to build the interpretable DCNN model. First, the authors design a BTSC module, in which each parent node generates two independent child layers, and then integrate them into a normal DCNN model. The main advantages of the BTSC are as follows: 1) child nodes of the different parent nodes do not interfere with each other; 2) parent and child nodes can inherit knowledge. Second, considering the activation level of neurons, the authors design an information coding objective to guide neural nodes to learn the particular information coding that is expected. Through the experiments, the authors can verify that: 1) the decision-making made by both the ResNet and DenseNet models can be explained well based on the "decision information flow path" (known as the decision-path) formed in the BTSC module; 2) the decision-path can reasonably interpret the decision reversal mechanism (Robustness mechanism) of the DCNN model; 3) the credibility of decision-making can be measured by the matching degree between the actual and expected decision-path.
期刊介绍:
CAAI Transactions on Intelligence Technology is a leading venue for original research on the theoretical and experimental aspects of artificial intelligence technology. We are a fully open access journal co-published by the Institution of Engineering and Technology (IET) and the Chinese Association for Artificial Intelligence (CAAI) providing research which is openly accessible to read and share worldwide.