{"title":"An Interpretable Deep Classifier for Counterfactual Generation","authors":"Wei Zhang, Brian Barr, J. Paisley","doi":"10.1145/3533271.3561722","DOIUrl":null,"url":null,"abstract":"Counterfactual explanation has been the core of interpretable machine learning, which requires a trained model to be able to not only infer but also justify its inference. This problem is crucial in many fields, such as fintech and the healthcare industry, where accurate decisions and their justifications are equally important. Many studies have leveraged the power of deep generative models for counterfactual generation. However, most focus on vision data and leave the latent space unsupervised. In this paper, we propose a new and general framework that uses a supervised extension to the Variational Auto-Encoder (VAE) with Normalizing Flow (NF) for simultaneous classification and counterfactual generation. We show experiments on two tabular financial data-sets, Lending Club (LCD) and Give Me Some Credit (GMC), which show that the model can achieve a state-of-art level prediction accuracy while also producing meaningful counterfactual examples to interpret and justify the classifier’s decision.","PeriodicalId":134888,"journal":{"name":"Proceedings of the Third ACM International Conference on AI in Finance","volume":"124 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the Third ACM International Conference on AI in Finance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3533271.3561722","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Counterfactual explanation has been the core of interpretable machine learning, which requires a trained model to be able to not only infer but also justify its inference. This problem is crucial in many fields, such as fintech and the healthcare industry, where accurate decisions and their justifications are equally important. Many studies have leveraged the power of deep generative models for counterfactual generation. However, most focus on vision data and leave the latent space unsupervised. In this paper, we propose a new and general framework that uses a supervised extension to the Variational Auto-Encoder (VAE) with Normalizing Flow (NF) for simultaneous classification and counterfactual generation. We show experiments on two tabular financial data-sets, Lending Club (LCD) and Give Me Some Credit (GMC), which show that the model can achieve a state-of-art level prediction accuracy while also producing meaningful counterfactual examples to interpret and justify the classifier’s decision.
反事实解释一直是可解释性机器学习的核心,这需要一个训练有素的模型不仅能够推断,而且能够证明其推断。这个问题在许多领域至关重要,比如金融科技和医疗保健行业,在这些领域,准确的决策及其理由同样重要。许多研究利用深度生成模型的力量来生成反事实。然而,大多数研究都集中在视觉数据上,留下了不受监督的潜在空间。在本文中,我们提出了一个新的和通用的框架,该框架使用了具有归一化流(NF)的变分自编码器(VAE)的监督扩展,用于同时分类和反事实生成。我们展示了两个表格金融数据集的实验,Lending Club (LCD)和Give Me Some Credit (GMC),这表明该模型可以达到最先进的预测精度,同时也产生了有意义的反事实示例来解释和证明分类器的决定。