Clinicians' roles and necessary levels of understanding in the use of artificial intelligence: A qualitative interview study with German medical students.

IF 3 1区 哲学 Q1 ETHICS BMC Medical Ethics Pub Date : 2024-10-07 DOI:10.1186/s12910-024-01109-w
F Funer, S Tinnemeyer, W Liedtke, S Salloch
{"title":"Clinicians' roles and necessary levels of understanding in the use of artificial intelligence: A qualitative interview study with German medical students.","authors":"F Funer, S Tinnemeyer, W Liedtke, S Salloch","doi":"10.1186/s12910-024-01109-w","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Artificial intelligence-driven Clinical Decision Support Systems (AI-CDSS) are being increasingly introduced into various domains of health care for diagnostic, prognostic, therapeutic and other purposes. A significant part of the discourse on ethically appropriate conditions relate to the levels of understanding and explicability needed for ensuring responsible clinical decision-making when using AI-CDSS. Empirical evidence on stakeholders' viewpoints on these issues is scarce so far. The present study complements the empirical-ethical body of research by, on the one hand, investigating the requirements for understanding and explicability in depth with regard to the rationale behind them. On the other hand, it surveys medical students at the end of their studies as stakeholders, of whom little data is available so far, but for whom AI-CDSS will be an important part of their medical practice.</p><p><strong>Methods: </strong>Fifteen semi-structured qualitative interviews (each lasting an average of 56 min) were conducted with German medical students to investigate their perspectives and attitudes on the use of AI-CDSS. The problem-centred interviews draw on two hypothetical case vignettes of AI-CDSS employed in nephrology and surgery. Interviewees' perceptions and convictions of their own clinical role and responsibilities in dealing with AI-CDSS were elicited as well as viewpoints on explicability as well as the necessary level of understanding and competencies needed on the clinicians' side. The qualitative data were analysed according to key principles of qualitative content analysis (Kuckartz).</p><p><strong>Results: </strong>In response to the central question about the necessary understanding of AI-CDSS tools and the emergence of their outputs as well as the reasons for the requirements placed on them, two types of argumentation could be differentiated inductively from the interviewees' statements: the first type, the clinician as a systemic trustee (or \"the one relying\"), highlights that there needs to be empirical evidence and adequate approval processes that guarantee minimised harm and a clinical benefit from the employment of an AI-CDSS. Based on proof of these requirements, the use of an AI-CDSS would be appropriate, as according to \"the one relying\", clinicians should choose those measures that statistically cause the least harm. The second type, the clinician as an individual expert (or \"the one controlling\"), sets higher prerequisites that go beyond ensuring empirical evidence and adequate approval processes. These higher prerequisites relate to the clinician's necessary level of competence and understanding of how a specific AI-CDSS works and how to use it properly in order to evaluate its outputs and to mitigate potential risks for the individual patient. Both types are unified in their high esteem of evidence-based clinical practice and the need to communicate with the patient on the use of medical AI. However, the interviewees' different conceptions of the clinician's role and responsibilities cause them to have different requirements regarding the clinician's understanding and explicability of an AI-CDSS beyond the proof of benefit.</p><p><strong>Conclusions: </strong>The study results highlight two different types among (future) clinicians regarding their view of the necessary levels of understanding and competence. These findings should inform the debate on appropriate training programmes and professional standards (e.g. clinical practice guidelines) that enable the safe and effective clinical employment of AI-CDSS in various clinical fields. While current approaches search for appropriate minimum requirements of the necessary understanding and competence, the differences between (future) clinicians in terms of their information and understanding needs described here can lead to more differentiated approaches to solutions.</p>","PeriodicalId":55348,"journal":{"name":"BMC Medical Ethics","volume":null,"pages":null},"PeriodicalIF":3.0000,"publicationDate":"2024-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11457475/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"BMC Medical Ethics","FirstCategoryId":"98","ListUrlMain":"https://doi.org/10.1186/s12910-024-01109-w","RegionNum":1,"RegionCategory":"哲学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ETHICS","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Artificial intelligence-driven Clinical Decision Support Systems (AI-CDSS) are being increasingly introduced into various domains of health care for diagnostic, prognostic, therapeutic and other purposes. A significant part of the discourse on ethically appropriate conditions relate to the levels of understanding and explicability needed for ensuring responsible clinical decision-making when using AI-CDSS. Empirical evidence on stakeholders' viewpoints on these issues is scarce so far. The present study complements the empirical-ethical body of research by, on the one hand, investigating the requirements for understanding and explicability in depth with regard to the rationale behind them. On the other hand, it surveys medical students at the end of their studies as stakeholders, of whom little data is available so far, but for whom AI-CDSS will be an important part of their medical practice.

Methods: Fifteen semi-structured qualitative interviews (each lasting an average of 56 min) were conducted with German medical students to investigate their perspectives and attitudes on the use of AI-CDSS. The problem-centred interviews draw on two hypothetical case vignettes of AI-CDSS employed in nephrology and surgery. Interviewees' perceptions and convictions of their own clinical role and responsibilities in dealing with AI-CDSS were elicited as well as viewpoints on explicability as well as the necessary level of understanding and competencies needed on the clinicians' side. The qualitative data were analysed according to key principles of qualitative content analysis (Kuckartz).

Results: In response to the central question about the necessary understanding of AI-CDSS tools and the emergence of their outputs as well as the reasons for the requirements placed on them, two types of argumentation could be differentiated inductively from the interviewees' statements: the first type, the clinician as a systemic trustee (or "the one relying"), highlights that there needs to be empirical evidence and adequate approval processes that guarantee minimised harm and a clinical benefit from the employment of an AI-CDSS. Based on proof of these requirements, the use of an AI-CDSS would be appropriate, as according to "the one relying", clinicians should choose those measures that statistically cause the least harm. The second type, the clinician as an individual expert (or "the one controlling"), sets higher prerequisites that go beyond ensuring empirical evidence and adequate approval processes. These higher prerequisites relate to the clinician's necessary level of competence and understanding of how a specific AI-CDSS works and how to use it properly in order to evaluate its outputs and to mitigate potential risks for the individual patient. Both types are unified in their high esteem of evidence-based clinical practice and the need to communicate with the patient on the use of medical AI. However, the interviewees' different conceptions of the clinician's role and responsibilities cause them to have different requirements regarding the clinician's understanding and explicability of an AI-CDSS beyond the proof of benefit.

Conclusions: The study results highlight two different types among (future) clinicians regarding their view of the necessary levels of understanding and competence. These findings should inform the debate on appropriate training programmes and professional standards (e.g. clinical practice guidelines) that enable the safe and effective clinical employment of AI-CDSS in various clinical fields. While current approaches search for appropriate minimum requirements of the necessary understanding and competence, the differences between (future) clinicians in terms of their information and understanding needs described here can lead to more differentiated approaches to solutions.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
临床医生在人工智能应用中的角色和必要的理解水平:对德国医科学生的定性访谈研究。
背景:人工智能驱动的临床决策支持系统(AI-CDSS)正被越来越多地引入医疗保健的各个领域,用于诊断、预后、治疗和其他目的。关于伦理适宜性条件的讨论,很大一部分涉及使用 AI-CDSS 时确保负责任的临床决策所需的理解程度和可解释性。迄今为止,有关利益相关者对这些问题看法的经验证据还很少。本研究一方面通过深入调查对理解和可解释性的要求及其背后的理论依据,对实证伦理研究进行了补充。另一方面,本研究还对即将毕业的医学生进行了调查,他们是研究的利益相关者,迄今为止,关于他们的数据很少,但对他们来说,AI-CDSS 将是其医疗实践的重要组成部分:对德国医科学生进行了 15 次半结构式定性访谈(每次平均 56 分钟),调查他们对使用 AI-CDSS 的看法和态度。以问题为中心的访谈以肾脏内科和外科使用人工智能数据采集系统的两个假设案例为基础。访谈内容包括受访者对自己在处理人工智能数据采集与分析系统时的临床角色和责任的看法和信念,以及对可解释性的观点和临床医生所需的理解和能力水平。根据定性内容分析的主要原则(Kuckartz)对定性数据进行了分析:在回答关于对人工智能 CDSS 工具的必要了解、其产出的出现以及对其提出要求的原因这一核心问题时,可以从受访者的陈述中归纳出两类论点:第一类是临床医生作为系统受托人(或 "依赖者"),强调需要有实证证据和适当的审批程序,以保证人工智能 CDSS 的使用能将危害降至最低并带来临床效益。根据 "依赖者 "的观点,临床医生应该选择那些从统计学角度来看伤害最小的措施。第二种类型是临床医生作为个人专家(或称 "控制者"),他们设定了更高的先决条件,这些先决条件超出了确保经验证据和适当审批程序的范围。这些更高的先决条件涉及临床医生的必要能力水平,以及对特定 AI-CDSS 如何工作和如何正确使用它的理解,以便评估其输出结果并降低对患者个人的潜在风险。这两类受访者都非常推崇循证临床实践,并认为有必要就医疗人工智能的使用与患者进行沟通。然而,受访者对临床医生角色和责任的不同理解,导致他们对临床医生对人工智能-CDSS的理解和可解释性有着不同的要求,而不仅仅是证明其益处:研究结果凸显了(未来)临床医生对必要的理解和能力水平的两种不同看法。这些研究结果应为有关适当的培训计划和专业标准(如临床实践指南)的讨论提供信息,以便在不同的临床领域安全有效地使用人工智能数据采集系统。虽然目前的方法是寻找适当的必要理解和能力的最低要求,但本文所描述的(未来)临床医生在信息和理解需求方面的差异可以导致更有区别的解决方案。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
BMC Medical Ethics
BMC Medical Ethics MEDICAL ETHICS-
CiteScore
5.20
自引率
7.40%
发文量
108
审稿时长
>12 weeks
期刊介绍: BMC Medical Ethics is an open access journal publishing original peer-reviewed research articles in relation to the ethical aspects of biomedical research and clinical practice, including professional choices and conduct, medical technologies, healthcare systems and health policies.
期刊最新文献
Public perceptions of the Hippocratic Oath in the U.K. 2023. Ethical challenges in organ transplantation for Syrian refugees in Türkiye. What ethical conflicts do internists in Spain, México and Argentina encounter? An international cross-sectional observational study based on a self-administrated survey. Medical futility at the end of life: the first qualitative study of ethical decision-making methods among Turkish doctors. Financial conflicts of interest among authors of clinical practice guideline for headache disorders in Japan.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1