A global scale comparison of risk aggregation in AI assessment frameworks

Anna Schmitz, Michael Mock, Rebekka Görge, Armin B. Cremers, Maximilian Poretschkin
{"title":"A global scale comparison of risk aggregation in AI assessment frameworks","authors":"Anna Schmitz,&nbsp;Michael Mock,&nbsp;Rebekka Görge,&nbsp;Armin B. Cremers,&nbsp;Maximilian Poretschkin","doi":"10.1007/s43681-024-00479-6","DOIUrl":null,"url":null,"abstract":"<div><p>AI applications bear inherent risks in various risk dimensions, such as insufficient reliability, robustness, fairness or data protection. It is well-known that trade-offs between these dimensions can arise, for example, a highly accurate AI application may reflect unfairness and bias of the real-world data, or may provide hard-to-explain outcomes because of its internal complexity. AI risk assessment frameworks aim to provide systematic approaches to risk assessment in various dimensions. The overall trustworthiness assessment is then generated by some form of risk aggregation among the risk dimensions. This paper provides a systematic overview on risk aggregation schemes used in existing AI risk assessment frameworks, focusing on the question how potential trade-offs among the risk dimensions are incorporated. To this end, we examine how the general risk notion, the application context, the extent of risk quantification, and specific instructions for evaluation may influence overall risk aggregation. We discuss our findings in the current frameworks in terms of whether they provide meaningful and practicable guidance. Lastly, we derive recommendations for the further operationalization of risk aggregation both from horizontal and vertical perspectives.</p></div>","PeriodicalId":72137,"journal":{"name":"AI and ethics","volume":"5 2","pages":"1407 - 1432"},"PeriodicalIF":0.0000,"publicationDate":"2024-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43681-024-00479-6.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"AI and ethics","FirstCategoryId":"1085","ListUrlMain":"https://link.springer.com/article/10.1007/s43681-024-00479-6","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

AI applications bear inherent risks in various risk dimensions, such as insufficient reliability, robustness, fairness or data protection. It is well-known that trade-offs between these dimensions can arise, for example, a highly accurate AI application may reflect unfairness and bias of the real-world data, or may provide hard-to-explain outcomes because of its internal complexity. AI risk assessment frameworks aim to provide systematic approaches to risk assessment in various dimensions. The overall trustworthiness assessment is then generated by some form of risk aggregation among the risk dimensions. This paper provides a systematic overview on risk aggregation schemes used in existing AI risk assessment frameworks, focusing on the question how potential trade-offs among the risk dimensions are incorporated. To this end, we examine how the general risk notion, the application context, the extent of risk quantification, and specific instructions for evaluation may influence overall risk aggregation. We discuss our findings in the current frameworks in terms of whether they provide meaningful and practicable guidance. Lastly, we derive recommendations for the further operationalization of risk aggregation both from horizontal and vertical perspectives.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
对人工智能评估框架中的风险汇总进行全球范围的比较
人工智能应用在各种风险维度上都存在固有风险,如可靠性、鲁棒性、公平性或数据保护不足。众所周知,这些维度之间的权衡可能会出现,例如,高度精确的人工智能应用程序可能反映现实世界数据的不公平和偏见,或者由于其内部复杂性而可能提供难以解释的结果。人工智能风险评估框架旨在为各个维度的风险评估提供系统的方法。然后通过风险维度之间的某种形式的风险聚合生成总体可信度评估。本文对现有人工智能风险评估框架中使用的风险聚合方案进行了系统概述,重点关注如何将风险维度之间的潜在权衡纳入其中。为此,我们检查一般风险概念、应用环境、风险量化的程度和评估的具体指示如何影响总体风险聚集。我们讨论我们的发现在目前的框架方面,他们是否提供有意义的和切实可行的指导。最后,我们从横向和纵向两方面提出了进一步实施风险聚合的建议。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Perceptions and predictors of trust in artificial intelligence use: ethical implications and regulatory oversight in Ghana Temporal authorship as a moral right: bounded inquiry in persistent AI delegation Machine Learning in Education Still unsafe: what’s holding us back on online safety for women When routing may bind: legitimacy conditions for typed jurisdiction allocation in AI-mediated governance
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1