M. Pazzani, Severine Soltani, Sateesh Kumar, Kamran Alipour, Aadil Ahamed
{"title":"Improving Explanations of Image Classifiers: Ensembles and Multitask Learning","authors":"M. Pazzani, Severine Soltani, Sateesh Kumar, Kamran Alipour, Aadil Ahamed","doi":"10.5121/ijaia.2022.13604","DOIUrl":null,"url":null,"abstract":"In explainable AI (XAI) for deep learning, saliency maps, heatmaps, or attention maps are commonly used to identify important regions for the classification of images of explanations. We address two important limitations of heatmaps. First, they do not correspond to type of explanations typically produced by human experts. Second, recent research has shown that many common XAI methods do not accurately identify the regions that human experts consider important. We propose using multitask learning to identify diagnostic features in images and averaging explanations from ensembles of learners to increase the accuracy of explanations. Our technique is general and can be used with multiple deep learning architectures and multiple XAI algorithms. We show that this method decreases the difference between regions of interest of XAI algorithms and those identified by human experts and the multitask learning supports the type of explanations produced by human experts. Furthermore, we show that human experts prefer the explanations produced by ensembles to those of individual networks.","PeriodicalId":391502,"journal":{"name":"International Journal of Artificial Intelligence & Applications","volume":"15 3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Artificial Intelligence & Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5121/ijaia.2022.13604","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In explainable AI (XAI) for deep learning, saliency maps, heatmaps, or attention maps are commonly used to identify important regions for the classification of images of explanations. We address two important limitations of heatmaps. First, they do not correspond to type of explanations typically produced by human experts. Second, recent research has shown that many common XAI methods do not accurately identify the regions that human experts consider important. We propose using multitask learning to identify diagnostic features in images and averaging explanations from ensembles of learners to increase the accuracy of explanations. Our technique is general and can be used with multiple deep learning architectures and multiple XAI algorithms. We show that this method decreases the difference between regions of interest of XAI algorithms and those identified by human experts and the multitask learning supports the type of explanations produced by human experts. Furthermore, we show that human experts prefer the explanations produced by ensembles to those of individual networks.