人工智能的可接受错误率是否应该低于人类?

BJR open Pub Date : 2023-01-01 DOI:10.1259/bjro.20220053
Anders Lenskjold, Janus Uhd Nybing, Charlotte Trampedach, Astrid Galsgaard, Mathias Willadsen Brejnebøl, Henriette Raaschou, Martin Høyer Rose, Mikael Boesen
{"title":"人工智能的可接受错误率是否应该低于人类?","authors":"Anders Lenskjold,&nbsp;Janus Uhd Nybing,&nbsp;Charlotte Trampedach,&nbsp;Astrid Galsgaard,&nbsp;Mathias Willadsen Brejnebøl,&nbsp;Henriette Raaschou,&nbsp;Martin Høyer Rose,&nbsp;Mikael Boesen","doi":"10.1259/bjro.20220053","DOIUrl":null,"url":null,"abstract":"<p><p>The first patient was misclassified in the diagnostic conclusion according to a local clinical expert opinion in a new clinical implementation of a knee osteoarthritis artificial intelligence (AI) algorithm at Bispebjerg-Frederiksberg University Hospital, Copenhagen, Denmark. In preparation for the evaluation of the AI algorithm, the implementation team collaborated with internal and external partners to plan workflows, and the algorithm was externally validated. After the misclassification, the team was left wondering: what is an acceptable error rate for a low-risk AI diagnostic algorithm? A survey among employees at the Department of Radiology showed significantly lower acceptable error rates for AI (6.8 %) than humans (11.3 %). A general mistrust of AI could cause the discrepancy in acceptable errors. AI may have the disadvantage of limited social capital and likeability compared to human co-workers, and therefore, less potential for forgiveness. Future AI development and implementation require further investigation of the fear of AI's unknown errors to enhance the trustworthiness of perceiving AI as a co-worker. Benchmark tools, transparency, and explainability are also needed to evaluate AI algorithms in clinical implementations to ensure acceptable performance.</p>","PeriodicalId":72419,"journal":{"name":"BJR open","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10301708/pdf/","citationCount":"1","resultStr":"{\"title\":\"Should artificial intelligence have lower acceptable error rates than humans?\",\"authors\":\"Anders Lenskjold,&nbsp;Janus Uhd Nybing,&nbsp;Charlotte Trampedach,&nbsp;Astrid Galsgaard,&nbsp;Mathias Willadsen Brejnebøl,&nbsp;Henriette Raaschou,&nbsp;Martin Høyer Rose,&nbsp;Mikael Boesen\",\"doi\":\"10.1259/bjro.20220053\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>The first patient was misclassified in the diagnostic conclusion according to a local clinical expert opinion in a new clinical implementation of a knee osteoarthritis artificial intelligence (AI) algorithm at Bispebjerg-Frederiksberg University Hospital, Copenhagen, Denmark. In preparation for the evaluation of the AI algorithm, the implementation team collaborated with internal and external partners to plan workflows, and the algorithm was externally validated. After the misclassification, the team was left wondering: what is an acceptable error rate for a low-risk AI diagnostic algorithm? A survey among employees at the Department of Radiology showed significantly lower acceptable error rates for AI (6.8 %) than humans (11.3 %). A general mistrust of AI could cause the discrepancy in acceptable errors. AI may have the disadvantage of limited social capital and likeability compared to human co-workers, and therefore, less potential for forgiveness. Future AI development and implementation require further investigation of the fear of AI's unknown errors to enhance the trustworthiness of perceiving AI as a co-worker. Benchmark tools, transparency, and explainability are also needed to evaluate AI algorithms in clinical implementations to ensure acceptable performance.</p>\",\"PeriodicalId\":72419,\"journal\":{\"name\":\"BJR open\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10301708/pdf/\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"BJR open\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1259/bjro.20220053\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"BJR open","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1259/bjro.20220053","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

在丹麦哥本哈根bispebjerge - frederiksberg大学医院,根据当地临床专家的意见,在膝关节骨关节炎人工智能(AI)算法的新临床实施中,第一位患者在诊断结论中被错误分类。在准备评估AI算法时,实施团队与内部和外部合作伙伴合作规划工作流程,并对算法进行外部验证。在错误分类之后,团队想知道:低风险人工智能诊断算法的可接受错误率是多少?一项针对放射科员工的调查显示,人工智能的可接受错误率(6.8%)明显低于人类(11.3%)。对人工智能的普遍不信任可能会导致可接受错误的差异。与人类同事相比,人工智能的缺点可能是社会资本和受欢迎程度有限,因此,获得宽恕的可能性更小。未来的人工智能开发和实施需要进一步研究对人工智能未知错误的恐惧,以提高将人工智能视为同事的可信度。还需要基准工具、透明度和可解释性来评估临床实施中的人工智能算法,以确保可接受的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

摘要图片

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Should artificial intelligence have lower acceptable error rates than humans?

The first patient was misclassified in the diagnostic conclusion according to a local clinical expert opinion in a new clinical implementation of a knee osteoarthritis artificial intelligence (AI) algorithm at Bispebjerg-Frederiksberg University Hospital, Copenhagen, Denmark. In preparation for the evaluation of the AI algorithm, the implementation team collaborated with internal and external partners to plan workflows, and the algorithm was externally validated. After the misclassification, the team was left wondering: what is an acceptable error rate for a low-risk AI diagnostic algorithm? A survey among employees at the Department of Radiology showed significantly lower acceptable error rates for AI (6.8 %) than humans (11.3 %). A general mistrust of AI could cause the discrepancy in acceptable errors. AI may have the disadvantage of limited social capital and likeability compared to human co-workers, and therefore, less potential for forgiveness. Future AI development and implementation require further investigation of the fear of AI's unknown errors to enhance the trustworthiness of perceiving AI as a co-worker. Benchmark tools, transparency, and explainability are also needed to evaluate AI algorithms in clinical implementations to ensure acceptable performance.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
审稿时长
18 weeks
期刊最新文献
Three-dimensional dose prediction based on deep convolutional neural networks for brain cancer in CyberKnife: accurate beam modelling of homogeneous tissue. Advancing radiology practice and research: harnessing the potential of large language models amidst imperfections. Improvement in paediatric CT use and justification: a single-centre experience. Deuterium MR spectroscopy: potential applications in oncology research. Unlocking the potential of photon counting detector CT for paediatric imaging: a pictorial essay.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1