Automated opioid risk scores: a case for machine learning-induced epistemic injustice in healthcare.

IF 3.4 2区 哲学 Q1 ETHICS Ethics and Information Technology Pub Date : 2023-01-01 DOI:10.1007/s10676-023-09676-z
Giorgia Pozzi
{"title":"Automated opioid risk scores: a case for machine learning-induced epistemic injustice in healthcare.","authors":"Giorgia Pozzi","doi":"10.1007/s10676-023-09676-z","DOIUrl":null,"url":null,"abstract":"<p><p>Artificial intelligence-based (AI) technologies such as machine learning (ML) systems are playing an increasingly relevant role in medicine and healthcare, bringing about novel ethical and epistemological issues that need to be timely addressed. Even though ethical questions connected to epistemic concerns have been at the center of the debate, it is going unnoticed how epistemic forms of injustice can be ML-induced, specifically in healthcare. I analyze the shortcomings of an ML system currently deployed in the USA to predict patients' likelihood of opioid addiction and misuse (PDMP algorithmic platforms). Drawing on this analysis, I aim to show that the wrong inflicted on epistemic agents involved in and affected by these systems' decision-making processes can be captured through the lenses of Miranda Fricker's account of <i>hermeneutical injustice</i>. I further argue that ML-induced hermeneutical injustice is particularly harmful due to what I define as an <i>automated hermeneutical appropriation</i> from the side of the ML system. The latter occurs if the ML system establishes meanings and shared hermeneutical resources without allowing for human oversight, impairing understanding and communication practices among stakeholders involved in medical decision-making. Furthermore and very much crucially, an automated hermeneutical appropriation can be recognized if physicians are strongly limited in their possibilities to safeguard patients from ML-induced hermeneutical injustice. Overall, my paper should expand the analysis of ethical issues raised by ML systems that are to be considered epistemic in nature, thus contributing to bridging the gap between these two dimensions in the ongoing debate.</p>","PeriodicalId":51495,"journal":{"name":"Ethics and Information Technology","volume":"25 1","pages":"3"},"PeriodicalIF":3.4000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9869303/pdf/","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ethics and Information Technology","FirstCategoryId":"91","ListUrlMain":"https://doi.org/10.1007/s10676-023-09676-z","RegionNum":2,"RegionCategory":"哲学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ETHICS","Score":null,"Total":0}
引用次数: 5

Abstract

Artificial intelligence-based (AI) technologies such as machine learning (ML) systems are playing an increasingly relevant role in medicine and healthcare, bringing about novel ethical and epistemological issues that need to be timely addressed. Even though ethical questions connected to epistemic concerns have been at the center of the debate, it is going unnoticed how epistemic forms of injustice can be ML-induced, specifically in healthcare. I analyze the shortcomings of an ML system currently deployed in the USA to predict patients' likelihood of opioid addiction and misuse (PDMP algorithmic platforms). Drawing on this analysis, I aim to show that the wrong inflicted on epistemic agents involved in and affected by these systems' decision-making processes can be captured through the lenses of Miranda Fricker's account of hermeneutical injustice. I further argue that ML-induced hermeneutical injustice is particularly harmful due to what I define as an automated hermeneutical appropriation from the side of the ML system. The latter occurs if the ML system establishes meanings and shared hermeneutical resources without allowing for human oversight, impairing understanding and communication practices among stakeholders involved in medical decision-making. Furthermore and very much crucially, an automated hermeneutical appropriation can be recognized if physicians are strongly limited in their possibilities to safeguard patients from ML-induced hermeneutical injustice. Overall, my paper should expand the analysis of ethical issues raised by ML systems that are to be considered epistemic in nature, thus contributing to bridging the gap between these two dimensions in the ongoing debate.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
自动阿片类药物风险评分:医疗保健中机器学习引起的认知不公正的案例。
基于人工智能(AI)的技术,如机器学习(ML)系统,在医学和医疗保健领域发挥着越来越重要的作用,带来了新的伦理和认识论问题,需要及时解决。尽管与认知相关的伦理问题一直是辩论的中心,但人们却没有注意到认知形式的不公正是如何被ml诱导的,特别是在医疗保健领域。我分析了目前在美国部署的ML系统的缺点,该系统用于预测患者阿片类药物成瘾和滥用的可能性(PDMP算法平台)。根据这一分析,我的目的是表明,通过米兰达·弗里克(Miranda Fricker)对解释学不公正的描述,可以捕捉到参与这些系统决策过程并受其影响的认知主体所遭受的错误。我进一步认为,机器学习导致的解释学不公正尤其有害,因为我将其定义为机器学习系统方面的自动解释学挪用。如果机器学习系统在不允许人类监督的情况下建立意义和共享解释学资源,则会发生后者,从而损害参与医疗决策的利益相关者之间的理解和沟通实践。此外,非常关键的是,如果医生在保护患者免受ml诱导的解释学不公正的可能性方面受到强烈限制,则可以识别自动解释学挪用。总的来说,我的论文应该扩展对机器学习系统提出的伦理问题的分析,这些问题本质上被认为是认识论的,从而有助于在正在进行的辩论中弥合这两个维度之间的差距。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
8.20
自引率
5.60%
发文量
46
期刊介绍: Ethics and Information Technology is a peer-reviewed journal dedicated to advancing the dialogue between moral philosophy and the field of information and communication technology (ICT). The journal aims to foster and promote reflection and analysis which is intended to make a constructive contribution to answering the ethical, social and political questions associated with the adoption, use, and development of ICT. Within the scope of the journal are also conceptual analysis and discussion of ethical ICT issues which arise in the context of technology assessment, cultural studies, public policy analysis and public administration, cognitive science, social and anthropological studies in technology, mass-communication, and legal studies.
期刊最新文献
Engineers on responsibility: feminist approaches to who’s responsible for ethical AI AI and the need for justification (to the patient). Trustworthiness of voting advice applications in Europe. Large language models and their big bullshit potential. How to teach responsible AI in Higher Education: challenges and opportunities
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1