Raising awareness of potential biases in medical machine learning: Experience from a Datathon.

Harry Hochheiser, Jesse Klug, Thomas Mathie, Tom J Pollard, Jesse D Raffa, Stephanie L Ballard, Evamarie A Conrad, Smitha Edakalavan, Allan Joseph, Nader Alnomasy, Sarah Nutman, Veronika Hill, Sumit Kapoor, Eddie Pérez Claudio, Olga V Kravchenko, Ruoting Li, Mehdi Nourelahi, Jenny Diaz, W Michael Taylor, Sydney R Rooney, Maeve Woeltje, Leo Anthony Celi, Christopher M Horvat
{"title":"Raising awareness of potential biases in medical machine learning: Experience from a Datathon.","authors":"Harry Hochheiser, Jesse Klug, Thomas Mathie, Tom J Pollard, Jesse D Raffa, Stephanie L Ballard, Evamarie A Conrad, Smitha Edakalavan, Allan Joseph, Nader Alnomasy, Sarah Nutman, Veronika Hill, Sumit Kapoor, Eddie Pérez Claudio, Olga V Kravchenko, Ruoting Li, Mehdi Nourelahi, Jenny Diaz, W Michael Taylor, Sydney R Rooney, Maeve Woeltje, Leo Anthony Celi, Christopher M Horvat","doi":"10.1101/2024.10.21.24315543","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>To challenge clinicians and informaticians to learn about potential sources of bias in medical machine learning models through investigation of data and predictions from an open-source severity of illness score.</p><p><strong>Methods: </strong>Over a two-day period (total elapsed time approximately 28 hours), we conducted a datathon that challenged interdisciplinary teams to investigate potential sources of bias in the Global Open Source Severity of Illness Score. Teams were invited to develop hypotheses, to use tools of their choosing to identify potential sources of bias, and to provide a final report.</p><p><strong>Results: </strong>Five teams participated, three of which included both informaticians and clinicians. Most (4/5) used Python for analyses, the remaining team used R. Common analysis themes included relationship of the GOSSIS-1 prediction score with demographics and care related variables; relationships between demographics and outcomes; calibration and factors related to the context of care; and the impact of missingness. Representativeness of the population, differences in calibration and model performance among groups, and differences in performance across hospital settings were identified as possible sources of bias.</p><p><strong>Discussion: </strong>Datathons are a promising approach for challenging developers and users to explore questions relating to unrecognized biases in medical machine learning algorithms.</p>","PeriodicalId":94281,"journal":{"name":"medRxiv : the preprint server for health sciences","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11537317/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"medRxiv : the preprint server for health sciences","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1101/2024.10.21.24315543","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Objective: To challenge clinicians and informaticians to learn about potential sources of bias in medical machine learning models through investigation of data and predictions from an open-source severity of illness score.

Methods: Over a two-day period (total elapsed time approximately 28 hours), we conducted a datathon that challenged interdisciplinary teams to investigate potential sources of bias in the Global Open Source Severity of Illness Score. Teams were invited to develop hypotheses, to use tools of their choosing to identify potential sources of bias, and to provide a final report.

Results: Five teams participated, three of which included both informaticians and clinicians. Most (4/5) used Python for analyses, the remaining team used R. Common analysis themes included relationship of the GOSSIS-1 prediction score with demographics and care related variables; relationships between demographics and outcomes; calibration and factors related to the context of care; and the impact of missingness. Representativeness of the population, differences in calibration and model performance among groups, and differences in performance across hospital settings were identified as possible sources of bias.

Discussion: Datathons are a promising approach for challenging developers and users to explore questions relating to unrecognized biases in medical machine learning algorithms.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
提高对医学机器学习中潜在偏见的认识:数据马拉松的经验。
目的挑战临床医生和信息学家,通过调查开源疾病严重程度评分的数据和预测,了解医学机器学习模型中潜在的偏差来源:在为期两天(总耗时约 28 小时)的时间里,我们开展了一次数据马拉松,挑战跨学科团队调查全球开源疾病严重程度评分中潜在的偏差来源。各小组应邀提出假设,使用自己选择的工具确定潜在的偏差来源,并提交最终报告:结果:共有五个团队参与,其中三个团队既有信息学家,也有临床医生。常见的分析主题包括 GOSSIS-1 预测得分与人口统计学和护理相关变量之间的关系;人口统计学与结果之间的关系;校准与护理环境相关因素;以及遗漏的影响。人口的代表性、组间校准和模型性能的差异以及不同医院环境下的性能差异被认为是可能的偏差来源:作者总结:弱势群体有可能受到有偏见的医疗机器学习模型的不利影响。为了避免这些不良后果,开发人员和用户必须了解识别潜在偏见所涉及的挑战。我们举办了一次数据马拉松,旨在让不同的参与者探索一个开源患者严重程度模型,找出潜在的偏见。五组临床医生和信息学家使用他们自己选择的工具来评估可能的偏差来源,应用了一系列分析技术并探索了多种特征。通过让不同参与者亲身体验有意义的数据,数据马拉松有可能提高人们对潜在偏见的认识,并促进开发公平公正的医疗机器学习模型的最佳实践。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
US Overdose Mortality Saw First Drop Below the Jalal-Burke Exponential Growth Curve in 2024. Suicidal thoughts and behaviours in Cape Town: a cross-sectional study of prevalence, social, contextual, and clinical correlates. Melanocyte loss dominates the vitiligo transcriptome: a rank-based meta-analysis. Fatigue Links Sociodemographic Risk to Pain Intensity and Spread in Two Surgical Cohorts. The Neuroendocrine Profile During the Trier Social Stress Test in College Freshmen Offers Insights into the Emergence of Anxiety and Depression Symptoms.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1