{"title":"破解密码:通过可解释模型增强对人工智能的信任","authors":"Vipin Gupta, Shailendra Shukla, Kumari Nikita","doi":"10.48047/resmil.v10i1.20","DOIUrl":null,"url":null,"abstract":"In this paper, we explore the critical challenges of building trust in artificial intelligence (AI) systems, particularly those characterized by black box models. The proliferation of complex and opaque AI models has raised concerns about a lack of interpretability, hindering users’ understanding and confidence in these systems Significant problem solved in this review addresses the importance of increasing the reliability of AI through semantic AI (XAI) approaches . clarify the complexity of the model To address this issue, our approach is a comprehensive review of the existing literature on XAI, black-box models, and their implications for reliability We thoroughly analyze various XAI methods, such as local interpretive model-agnostic explanations (LIME), SHapley explanatory agnostic explanations (SHAP). and reflection methods, in addition to clarifying their efforts aimed at making AI models transparent, we examine real-world case studies in which the use of XAI has enhanced trustworthiness of AI systems have improved in various sectors. The main findings of our study highlight the important role of XAI in reducing the uncertainty associated with black-box models. We highlight examples where the adoption of interpretable approaches not only increased the interpretability of AI systems but also enhanced user confidence. By providing transparent insights into decision-making processes, XAI is proving to help remove complex models and establish a foundation of trust between users and AI systems The implications of our research apply to a range of industries that rely on AI, including healthcare, finance and autonomous systems. While opening up the benefits of XAI for building trust, we recommend its inclusion in AI development practices and highlight possible future developments in this area. However, our study acknowledges the existing","PeriodicalId":517991,"journal":{"name":"resmilitaris","volume":"91 5","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Cracking the Code: Enhancing Trust in AI through Explainable Models\",\"authors\":\"Vipin Gupta, Shailendra Shukla, Kumari Nikita\",\"doi\":\"10.48047/resmil.v10i1.20\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we explore the critical challenges of building trust in artificial intelligence (AI) systems, particularly those characterized by black box models. The proliferation of complex and opaque AI models has raised concerns about a lack of interpretability, hindering users’ understanding and confidence in these systems Significant problem solved in this review addresses the importance of increasing the reliability of AI through semantic AI (XAI) approaches . clarify the complexity of the model To address this issue, our approach is a comprehensive review of the existing literature on XAI, black-box models, and their implications for reliability We thoroughly analyze various XAI methods, such as local interpretive model-agnostic explanations (LIME), SHapley explanatory agnostic explanations (SHAP). and reflection methods, in addition to clarifying their efforts aimed at making AI models transparent, we examine real-world case studies in which the use of XAI has enhanced trustworthiness of AI systems have improved in various sectors. The main findings of our study highlight the important role of XAI in reducing the uncertainty associated with black-box models. We highlight examples where the adoption of interpretable approaches not only increased the interpretability of AI systems but also enhanced user confidence. By providing transparent insights into decision-making processes, XAI is proving to help remove complex models and establish a foundation of trust between users and AI systems The implications of our research apply to a range of industries that rely on AI, including healthcare, finance and autonomous systems. While opening up the benefits of XAI for building trust, we recommend its inclusion in AI development practices and highlight possible future developments in this area. However, our study acknowledges the existing\",\"PeriodicalId\":517991,\"journal\":{\"name\":\"resmilitaris\",\"volume\":\"91 5\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"resmilitaris\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.48047/resmil.v10i1.20\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"resmilitaris","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48047/resmil.v10i1.20","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Cracking the Code: Enhancing Trust in AI through Explainable Models
In this paper, we explore the critical challenges of building trust in artificial intelligence (AI) systems, particularly those characterized by black box models. The proliferation of complex and opaque AI models has raised concerns about a lack of interpretability, hindering users’ understanding and confidence in these systems Significant problem solved in this review addresses the importance of increasing the reliability of AI through semantic AI (XAI) approaches . clarify the complexity of the model To address this issue, our approach is a comprehensive review of the existing literature on XAI, black-box models, and their implications for reliability We thoroughly analyze various XAI methods, such as local interpretive model-agnostic explanations (LIME), SHapley explanatory agnostic explanations (SHAP). and reflection methods, in addition to clarifying their efforts aimed at making AI models transparent, we examine real-world case studies in which the use of XAI has enhanced trustworthiness of AI systems have improved in various sectors. The main findings of our study highlight the important role of XAI in reducing the uncertainty associated with black-box models. We highlight examples where the adoption of interpretable approaches not only increased the interpretability of AI systems but also enhanced user confidence. By providing transparent insights into decision-making processes, XAI is proving to help remove complex models and establish a foundation of trust between users and AI systems The implications of our research apply to a range of industries that rely on AI, including healthcare, finance and autonomous systems. While opening up the benefits of XAI for building trust, we recommend its inclusion in AI development practices and highlight possible future developments in this area. However, our study acknowledges the existing