{"title":"利用特征测量 XAI 方法生成的解释。","authors":"Cătălin-Mihai Pesecan, Lăcrămioara Stoicu-Tivadar","doi":"10.3233/SHTI241102","DOIUrl":null,"url":null,"abstract":"<p><p>An increasing number of explainability methods began to emerge as a response for the black-box methods used to make decisions that could not be easily explained. This created the need for a better evaluation for these methods. In this paper we propose a new method for evaluation based on features. The main advantage of applying the proposed method to CNNs explanations are: a fully automated way to measure the quality of an explanation and the fact that the score uses the same information as the CNN, in this way being able to offer a measure of the quality of explanation that can be obtained automatically, ensuring that the human bias will not be present in the measurement of the explanation.</p>","PeriodicalId":94357,"journal":{"name":"Studies in health technology and informatics","volume":"321 ","pages":"250-253"},"PeriodicalIF":0.0000,"publicationDate":"2024-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Measurement of Explanations Generated by XAI Methods Using Features.\",\"authors\":\"Cătălin-Mihai Pesecan, Lăcrămioara Stoicu-Tivadar\",\"doi\":\"10.3233/SHTI241102\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>An increasing number of explainability methods began to emerge as a response for the black-box methods used to make decisions that could not be easily explained. This created the need for a better evaluation for these methods. In this paper we propose a new method for evaluation based on features. The main advantage of applying the proposed method to CNNs explanations are: a fully automated way to measure the quality of an explanation and the fact that the score uses the same information as the CNN, in this way being able to offer a measure of the quality of explanation that can be obtained automatically, ensuring that the human bias will not be present in the measurement of the explanation.</p>\",\"PeriodicalId\":94357,\"journal\":{\"name\":\"Studies in health technology and informatics\",\"volume\":\"321 \",\"pages\":\"250-253\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-11-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Studies in health technology and informatics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3233/SHTI241102\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Studies in health technology and informatics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3233/SHTI241102","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Measurement of Explanations Generated by XAI Methods Using Features.
An increasing number of explainability methods began to emerge as a response for the black-box methods used to make decisions that could not be easily explained. This created the need for a better evaluation for these methods. In this paper we propose a new method for evaluation based on features. The main advantage of applying the proposed method to CNNs explanations are: a fully automated way to measure the quality of an explanation and the fact that the score uses the same information as the CNN, in this way being able to offer a measure of the quality of explanation that can be obtained automatically, ensuring that the human bias will not be present in the measurement of the explanation.