破解密码:通过可解释模型增强对人工智能的信任

Vipin Gupta, Shailendra Shukla, Kumari Nikita
{"title":"破解密码:通过可解释模型增强对人工智能的信任","authors":"Vipin Gupta, Shailendra Shukla, Kumari Nikita","doi":"10.48047/resmil.v10i1.20","DOIUrl":null,"url":null,"abstract":"In this paper, we explore the critical challenges of building trust in artificial intelligence (AI) systems, particularly those characterized by black box models. The proliferation of complex and opaque AI models has raised concerns about a lack of interpretability, hindering users’ understanding and confidence in these systems Significant problem solved in this review addresses the importance of increasing the reliability of AI through semantic AI (XAI) approaches . clarify the complexity of the model To address this issue, our approach is a comprehensive review of the existing literature on XAI, black-box models, and their implications for reliability We thoroughly analyze various XAI methods, such as local interpretive model-agnostic explanations (LIME), SHapley explanatory agnostic explanations (SHAP). and reflection methods, in addition to clarifying their efforts aimed at making AI models transparent, we examine real-world case studies in which the use of XAI has enhanced trustworthiness of AI systems have improved in various sectors. The main findings of our study highlight the important role of XAI in reducing the uncertainty associated with black-box models. We highlight examples where the adoption of interpretable approaches not only increased the interpretability of AI systems but also enhanced user confidence. By providing transparent insights into decision-making processes, XAI is proving to help remove complex models and establish a foundation of trust between users and AI systems The implications of our research apply to a range of industries that rely on AI, including healthcare, finance and autonomous systems. While opening up the benefits of XAI for building trust, we recommend its inclusion in AI development practices and highlight possible future developments in this area. However, our study acknowledges the existing","PeriodicalId":517991,"journal":{"name":"resmilitaris","volume":"91 5","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Cracking the Code: Enhancing Trust in AI through Explainable Models\",\"authors\":\"Vipin Gupta, Shailendra Shukla, Kumari Nikita\",\"doi\":\"10.48047/resmil.v10i1.20\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we explore the critical challenges of building trust in artificial intelligence (AI) systems, particularly those characterized by black box models. The proliferation of complex and opaque AI models has raised concerns about a lack of interpretability, hindering users’ understanding and confidence in these systems Significant problem solved in this review addresses the importance of increasing the reliability of AI through semantic AI (XAI) approaches . clarify the complexity of the model To address this issue, our approach is a comprehensive review of the existing literature on XAI, black-box models, and their implications for reliability We thoroughly analyze various XAI methods, such as local interpretive model-agnostic explanations (LIME), SHapley explanatory agnostic explanations (SHAP). and reflection methods, in addition to clarifying their efforts aimed at making AI models transparent, we examine real-world case studies in which the use of XAI has enhanced trustworthiness of AI systems have improved in various sectors. The main findings of our study highlight the important role of XAI in reducing the uncertainty associated with black-box models. We highlight examples where the adoption of interpretable approaches not only increased the interpretability of AI systems but also enhanced user confidence. By providing transparent insights into decision-making processes, XAI is proving to help remove complex models and establish a foundation of trust between users and AI systems The implications of our research apply to a range of industries that rely on AI, including healthcare, finance and autonomous systems. While opening up the benefits of XAI for building trust, we recommend its inclusion in AI development practices and highlight possible future developments in this area. However, our study acknowledges the existing\",\"PeriodicalId\":517991,\"journal\":{\"name\":\"resmilitaris\",\"volume\":\"91 5\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"resmilitaris\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.48047/resmil.v10i1.20\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"resmilitaris","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48047/resmil.v10i1.20","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

在本文中,我们探讨了在人工智能(AI)系统中建立信任的关键挑战,尤其是那些以黑盒模型为特征的系统。复杂而不透明的人工智能模型的激增引发了人们对缺乏可解释性的担忧,阻碍了用户对这些系统的理解和信心。 本综述解决的重要问题是通过语义人工智能(XAI)方法提高人工智能可靠性的重要性。我们深入分析了各种 XAI 方法,如本地解释性模型不可知论解释(LIME)、SHapley 解释性不可知论解释(SHAP)和反思方法,除了阐明它们旨在使人工智能模型透明化的努力之外,我们还研究了现实世界中的案例研究,在这些案例研究中,XAI 的使用提高了各行业人工智能系统的可信度。我们研究的主要发现强调了 XAI 在减少与黑箱模型相关的不确定性方面的重要作用。我们着重举例说明,采用可解释方法不仅提高了人工智能系统的可解释性,还增强了用户信心。事实证明,通过提供对决策过程的透明见解,XAI 有助于消除复杂的模型,并在用户和人工智能系统之间建立信任基础。 我们研究的意义适用于一系列依赖人工智能的行业,包括医疗保健、金融和自主系统。在开放 XAI 对建立信任的益处的同时,我们建议将其纳入人工智能开发实践,并强调了该领域未来可能的发展。不过,我们的研究也承认现有的
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Cracking the Code: Enhancing Trust in AI through Explainable Models
In this paper, we explore the critical challenges of building trust in artificial intelligence (AI) systems, particularly those characterized by black box models. The proliferation of complex and opaque AI models has raised concerns about a lack of interpretability, hindering users’ understanding and confidence in these systems Significant problem solved in this review addresses the importance of increasing the reliability of AI through semantic AI (XAI) approaches . clarify the complexity of the model To address this issue, our approach is a comprehensive review of the existing literature on XAI, black-box models, and their implications for reliability We thoroughly analyze various XAI methods, such as local interpretive model-agnostic explanations (LIME), SHapley explanatory agnostic explanations (SHAP). and reflection methods, in addition to clarifying their efforts aimed at making AI models transparent, we examine real-world case studies in which the use of XAI has enhanced trustworthiness of AI systems have improved in various sectors. The main findings of our study highlight the important role of XAI in reducing the uncertainty associated with black-box models. We highlight examples where the adoption of interpretable approaches not only increased the interpretability of AI systems but also enhanced user confidence. By providing transparent insights into decision-making processes, XAI is proving to help remove complex models and establish a foundation of trust between users and AI systems The implications of our research apply to a range of industries that rely on AI, including healthcare, finance and autonomous systems. While opening up the benefits of XAI for building trust, we recommend its inclusion in AI development practices and highlight possible future developments in this area. However, our study acknowledges the existing
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
A Production inventory model for deteriorating items with quadratic time dependent demand with two levels of Production OPTIMIZED ADAPTIVE NOISECANCELLATION ON FPGA India's Evolving Climate Diplomacy: Reassessing the Present, Envisioning the Future Deciphering the Influence of PMJDY on Financial Inclusion and Development Sustainability The Impact of COVID-19 on the Nightlife Industry in North Cyprus
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1