{"title":"iGOS++: integrated gradient optimized saliency by bilateral perturbations","authors":"S. Khorram, T. Lawson, Fuxin Li","doi":"10.1145/3450439.3451865","DOIUrl":null,"url":null,"abstract":"The black-box nature of the deep networks makes the explanation for \"why\" they make certain predictions extremely challenging. Saliency maps are one of the most widely-used local explanation tools to alleviate this problem. One of the primary approaches for generating saliency maps is by optimizing for a mask over the input dimensions so that the output of the network for a given class is influenced the most. However, prior work only studies such influence by removing evidence from the input. In this paper, we present iGOS++, a framework to generate saliency maps for blackbox networks by considering both removal and preservation of evidence. Additionally, we introduce the bilateral total variation term to the optimization that improves the continuity of the saliency map especially under high resolution and with thin object parts. We validate the capabilities of iGOS++ by extensive experiments and comparison against state-of-the-art saliency map methods. Our results show significant improvement in locating salient regions that are directly interpretable by humans. Besides, we showcased the capabilities of our method, iGOS++, in a real-world application of AI on medical data: the task of classifying COVID-19 cases from x-ray images. To our surprise, we discovered that sometimes the classifier is overfitted to the text characters printed on the x-ray images when performing classification rather than focusing on the evidence in the lungs. Fixing this overfitting issue by data cleansing significantly improved the precision and recall of the classifier.","PeriodicalId":87342,"journal":{"name":"Proceedings of the ACM Conference on Health, Inference, and Learning","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"15","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the ACM Conference on Health, Inference, and Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3450439.3451865","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 15
Abstract
The black-box nature of the deep networks makes the explanation for "why" they make certain predictions extremely challenging. Saliency maps are one of the most widely-used local explanation tools to alleviate this problem. One of the primary approaches for generating saliency maps is by optimizing for a mask over the input dimensions so that the output of the network for a given class is influenced the most. However, prior work only studies such influence by removing evidence from the input. In this paper, we present iGOS++, a framework to generate saliency maps for blackbox networks by considering both removal and preservation of evidence. Additionally, we introduce the bilateral total variation term to the optimization that improves the continuity of the saliency map especially under high resolution and with thin object parts. We validate the capabilities of iGOS++ by extensive experiments and comparison against state-of-the-art saliency map methods. Our results show significant improvement in locating salient regions that are directly interpretable by humans. Besides, we showcased the capabilities of our method, iGOS++, in a real-world application of AI on medical data: the task of classifying COVID-19 cases from x-ray images. To our surprise, we discovered that sometimes the classifier is overfitted to the text characters printed on the x-ray images when performing classification rather than focusing on the evidence in the lungs. Fixing this overfitting issue by data cleansing significantly improved the precision and recall of the classifier.