首页 > 最新文献

JAMIA Open最新文献

英文 中文
Assessing the quality of reporting in artificial intelligence/machine learning research for cardiac amyloidosis. 评估心脏淀粉样变性人工智能/机器学习研究报告的质量。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-08 eCollection Date: 2025-10-01 DOI: 10.1093/jamiaopen/ooaf104
Asiful Arefeen, Simar Singh, Crystal Razavi, Hassan Ghasemzadeh, Sandesh Dev

Objectives: Despite the rapid development of AI in clinical medicine, reproducibility and methodological limitations hinder its clinical utility. In response, MINimum Information for Medical AI Reporting (MINIMAR) standards were introduced to enhance publication standards and reduce bias, but their application remains unexplored. In this review, we sought to assesses the quality of reporting in AI/ML studies of cardiac amyloidosis (CA) an increasingly important cause of heart failure.

Materials and methods: Using PRISMA-ScR guidelines, we performed a scoping review of English-language articles published through May 2023 which applied AI/ML techniques to diagnose or predict CA. Non-CA studies and those with selective feature sets were excluded. Two researchers independently screened and extracted data. In all, 20 studies met criteria and were assessed for adherence to MINIMAR standards.

Results: The studies showed variable compliance with MINIMAR. Most reported participant age (90%) and gender (85%), but only 25% included ethnic or racial data, and none provided socioeconomic details. The majority (95%) developed diagnostic models, yet only 85% clearly described training features, and 20% addressed missing data. Model evaluation revealed gaps; 80% reported internal validation, but only 20% conducted external validation.

Discussion and conclusion: This study, one of the first to apply MINIMAR criteria to ML research in CA, reveals significant variability and deficiencies in reporting, particularly in patient demographics, model architecture, and evaluation. These findings underscore the need for stricter adherence to standardized reporting guidelines to enhance the reliability, generalizability, and clinical applicability of ML/AI models in CA.

目的:尽管人工智能在临床医学中的发展迅速,但可重复性和方法学的局限性阻碍了其临床应用。为此,引入了医疗人工智能报告的最低信息(MINIMAR)标准,以提高出版标准并减少偏见,但其应用仍未探索。在这篇综述中,我们试图评估心脏淀粉样变性(CA)的AI/ML研究报告的质量,这是心力衰竭日益重要的原因。材料和方法:使用PRISMA-ScR指南,我们对截至2023年5月发表的应用AI/ML技术诊断或预测CA的英文文章进行了范围审查。非CA研究和具有选择性特征集的研究被排除在外。两名研究人员独立筛选和提取数据。总共有20项研究符合标准,并对遵守MINIMAR标准进行了评估。结果:研究显示MINIMAR的依从性不同。大多数报告的参与者年龄(90%)和性别(85%),但只有25%包括民族或种族数据,没有人提供社会经济细节。大多数(95%)开发了诊断模型,但只有85%明确描述了训练特征,20%解决了缺失数据。模型评价揭示了差距;80%报告了内部验证,但只有20%进行了外部验证。讨论和结论:本研究是第一个将MINIMAR标准应用于CA ML研究的研究之一,它揭示了报告的显著差异和缺陷,特别是在患者人口统计学、模型架构和评估方面。这些发现强调需要更严格地遵守标准化报告指南,以提高CA中ML/AI模型的可靠性、普遍性和临床适用性。
{"title":"Assessing the quality of reporting in artificial intelligence/machine learning research for cardiac amyloidosis.","authors":"Asiful Arefeen, Simar Singh, Crystal Razavi, Hassan Ghasemzadeh, Sandesh Dev","doi":"10.1093/jamiaopen/ooaf104","DOIUrl":"10.1093/jamiaopen/ooaf104","url":null,"abstract":"<p><strong>Objectives: </strong>Despite the rapid development of AI in clinical medicine, reproducibility and methodological limitations hinder its clinical utility. In response, MINimum Information for Medical AI Reporting (MINIMAR) standards were introduced to enhance publication standards and reduce bias, but their application remains unexplored. In this review, we sought to assesses the quality of reporting in AI/ML studies of cardiac amyloidosis (CA) an increasingly important cause of heart failure.</p><p><strong>Materials and methods: </strong>Using PRISMA-ScR guidelines, we performed a scoping review of English-language articles published through May 2023 which applied AI/ML techniques to diagnose or predict CA. Non-CA studies and those with selective feature sets were excluded. Two researchers independently screened and extracted data. In all, 20 studies met criteria and were assessed for adherence to MINIMAR standards.</p><p><strong>Results: </strong>The studies showed variable compliance with MINIMAR. Most reported participant age (90%) and gender (85%), but only 25% included ethnic or racial data, and none provided socioeconomic details. The majority (95%) developed diagnostic models, yet only 85% clearly described training features, and 20% addressed missing data. Model evaluation revealed gaps; 80% reported internal validation, but only 20% conducted external validation.</p><p><strong>Discussion and conclusion: </strong>This study, one of the first to apply MINIMAR criteria to ML research in CA, reveals significant variability and deficiencies in reporting, particularly in patient demographics, model architecture, and evaluation. These findings underscore the need for stricter adherence to standardized reporting guidelines to enhance the reliability, generalizability, and clinical applicability of ML/AI models in CA.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf104"},"PeriodicalIF":3.4,"publicationDate":"2025-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12507469/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145259548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting falls using electronic health records: a time series approach. 使用电子健康记录预测跌倒:时间序列方法。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-03 eCollection Date: 2025-10-01 DOI: 10.1093/jamiaopen/ooaf116
Peter J Hoover, Terri L Blumke, Anna D Ware, Malvika Pillai, Zachary P Veigulis, Catherine M Curtin, Thomas F Osborne

Objective: To develop a more accurate fall prediction model within the Veterans Health Administration.

Materials and methods: The cohort included Veterans admitted to a Veterans Health Administration acute care setting from July 1, 2020, to June 30, 2022, with a length of stay between 1 and 7 days. Demographic and clinical data were obtained through electronic health records. Veterans were identified as having a documented fall through clinical progress notes. A transformer model was used to obtain features of this data, which was then used to train a Light Gradient-Boosting Machine for classification and prediction. Area under the precision-recall curve assisted in model tuning, with geometric mean used to define an optimal classification threshold.

Results: Among 242,844 Veterans assessed, 5965 (2.5%) were documented as having a fall during their clinical stay. Employing a transformer model with a Light Gradient-Boosting Machine resulted in an area under the curve of .851 and an area under the precision-recall curve of .285. With an accuracy of 76.3%, the model resulted in a specificity of 76.2% and a sensitivity of 77.3%.

Discussion: Prior evaluations have highlighted limitations of the Morse Fall Scale (MFS) in accurately assessing fall risk. Developing a time series classification model using existing electronic health record data, our model outperformed traditional MFS-based evaluations and other fall-risk models. Future work is necessary to address limitations, including class imbalance and the need for prospective validation.

Conclusion: An improvement over the MFS, this model, automatically calculated from existing data, can provide a more efficient and accurate means for identifying patients at risk of fall.

目的:为退伍军人健康管理局建立更准确的跌倒预测模型。材料和方法:该队列包括2020年7月1日至2022年6月30日在退伍军人健康管理局急症护理机构住院的退伍军人,住院时间在1至7天之间。通过电子健康记录获得人口统计和临床数据。通过临床进展记录,退伍军人被确定为有跌倒记录。利用变压器模型获取该数据的特征,然后利用该特征训练Light Gradient-Boosting Machine进行分类和预测。准确率-召回率曲线下的面积有助于模型调整,几何平均值用于定义最佳分类阈值。结果:在接受评估的242,844名退伍军人中,5965名(2.5%)被记录为在临床住院期间跌倒。采用带光梯度增强机的变压器模型,得到曲线下的面积。精密度-召回曲线下面积为0.285。该模型的准确率为76.3%,特异性为76.2%,敏感性为77.3%。讨论:先前的评估强调了莫尔斯坠落量表(MFS)在准确评估坠落风险方面的局限性。利用现有的电子健康记录数据开发了一个时间序列分类模型,我们的模型优于传统的基于mfs的评估和其他跌倒风险模型。未来的工作需要解决局限性,包括类别不平衡和前瞻性验证的需要。结论:该模型是对MFS的改进,可根据现有数据自动计算,为识别有跌倒风险的患者提供更有效和准确的方法。
{"title":"Predicting falls using electronic health records: a time series approach.","authors":"Peter J Hoover, Terri L Blumke, Anna D Ware, Malvika Pillai, Zachary P Veigulis, Catherine M Curtin, Thomas F Osborne","doi":"10.1093/jamiaopen/ooaf116","DOIUrl":"10.1093/jamiaopen/ooaf116","url":null,"abstract":"<p><strong>Objective: </strong>To develop a more accurate fall prediction model within the Veterans Health Administration.</p><p><strong>Materials and methods: </strong>The cohort included Veterans admitted to a Veterans Health Administration acute care setting from July 1, 2020, to June 30, 2022, with a length of stay between 1 and 7 days. Demographic and clinical data were obtained through electronic health records. Veterans were identified as having a documented fall through clinical progress notes. A transformer model was used to obtain features of this data, which was then used to train a Light Gradient-Boosting Machine for classification and prediction. Area under the precision-recall curve assisted in model tuning, with geometric mean used to define an optimal classification threshold.</p><p><strong>Results: </strong>Among 242,844 Veterans assessed, 5965 (2.5%) were documented as having a fall during their clinical stay. Employing a transformer model with a Light Gradient-Boosting Machine resulted in an area under the curve of .851 and an area under the precision-recall curve of .285. With an accuracy of 76.3%, the model resulted in a specificity of 76.2% and a sensitivity of 77.3%.</p><p><strong>Discussion: </strong>Prior evaluations have highlighted limitations of the Morse Fall Scale (MFS) in accurately assessing fall risk. Developing a time series classification model using existing electronic health record data, our model outperformed traditional MFS-based evaluations and other fall-risk models. Future work is necessary to address limitations, including class imbalance and the need for prospective validation.</p><p><strong>Conclusion: </strong>An improvement over the MFS, this model, automatically calculated from existing data, can provide a more efficient and accurate means for identifying patients at risk of fall.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf116"},"PeriodicalIF":3.4,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12492486/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145233301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analyzing the impact of pharmacogenomics-guided nonsteroidal anti-inflammatory drug alerts in clinical practice. 分析药物基因组学指导的非甾体抗炎药警报在临床实践中的影响。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-03 eCollection Date: 2025-10-01 DOI: 10.1093/jamiaopen/ooaf112
Amanda Massmann, Natasha J Petry, Max Weaver, Halle Brady, Roxana A Lupu

Objectives: This study evaluates response rates of pharmacogenomics (PGx) nonsteroidal anti-inflammatory drugs (NSAIDs) clinical decision support (CDS) alerts at Sanford Health from May 2020 to December 2024.

Materials and methods: A retrospective analysis was conducted on PGx NSAIDs interruptive alerts. Response options were classified into five categories (1) continuation of triggering NSAID order, (2) dose modification, (3) alternative NSAID ordered without PGx implications, (4) alternative analgesic (ie, opioid) ordered, and (5) discontinuation of NSAID without alternative therapy.

Results: The study analyzed 2361 alert instances from 978 patients. The most common response was discontinuing NSAID without alternative therapy (43%). Dose modifications and orders for alternative analgesics comprised 2.57% and 14.67% of responses, respectively. The initial acceptance rate was 62.6%. Prior NSAID use significantly impacted override rates (60% vs 40%, P < .001). A 409-day breaking point was observed to affect alert acceptance rates, with the highest acceptance in NSAID naïve patients (96.1%).

Discussion: PGx NSAIDs CDS alert acceptance rates were higher compared to general CDS acceptance rates. This study highlights opportunities for continuous improvement including optimizing alert modality, modifying alert criteria to include look-back periods, and implementing genetically adapted ordersets.

Conclusion: The initial acceptance rate of PGx NSAIDs CDS alerts was 62.6%, however, with significantly higher acceptance rates in NSAID naïve patients (62.6% vs 96.1%, P < .001). Integration of CDS is vital to the successful implementation of PGx in clinical practice.

目的:本研究评估2020年5月至2024年12月Sanford Health的药物基因组学(PGx)非甾体抗炎药(NSAIDs)临床决策支持(CDS)警报的响应率。材料和方法:回顾性分析PGx非甾体抗炎药中断警报。反应选择分为五类(1)继续触发NSAID订单,(2)剂量调整,(3)无PGx影响的替代NSAID订单,(4)替代镇痛药(如阿片类药物)订单,(5)无替代治疗的NSAID停药。结果:本研究分析了978例患者的2361例报警病例。最常见的反应是停止非甾体抗炎药而不进行替代治疗(43%)。剂量调整和替代镇痛药的顺序分别占应答的2.57%和14.67%。初始录取率为62.6%。先前使用非甾体抗炎药显著影响覆盖率(60% vs 40%, P)讨论:PGx非甾体抗炎药CDS警报接受率高于一般CDS接受率。本研究强调了持续改进的机会,包括优化警报模式,修改警报标准以包括回顾期,以及实现基因适应的订单集。结论:PGx非甾体抗炎药CDS预警的初始接受率为62.6%,而NSAID naïve患者的接受率明显更高(62.6% vs 96.1%, P
{"title":"Analyzing the impact of pharmacogenomics-guided nonsteroidal anti-inflammatory drug alerts in clinical practice.","authors":"Amanda Massmann, Natasha J Petry, Max Weaver, Halle Brady, Roxana A Lupu","doi":"10.1093/jamiaopen/ooaf112","DOIUrl":"10.1093/jamiaopen/ooaf112","url":null,"abstract":"<p><strong>Objectives: </strong>This study evaluates response rates of pharmacogenomics (PGx) nonsteroidal anti-inflammatory drugs (NSAIDs) clinical decision support (CDS) alerts at Sanford Health from May 2020 to December 2024.</p><p><strong>Materials and methods: </strong>A retrospective analysis was conducted on PGx NSAIDs interruptive alerts. Response options were classified into five categories (1) continuation of triggering NSAID order, (2) dose modification, (3) alternative NSAID ordered without PGx implications, (4) alternative analgesic (ie, opioid) ordered, and (5) discontinuation of NSAID without alternative therapy.</p><p><strong>Results: </strong>The study analyzed 2361 alert instances from 978 patients. The most common response was discontinuing NSAID without alternative therapy (43%). Dose modifications and orders for alternative analgesics comprised 2.57% and 14.67% of responses, respectively. The initial acceptance rate was 62.6%. Prior NSAID use significantly impacted override rates (60% vs 40%, <i>P</i> < .001). A 409-day breaking point was observed to affect alert acceptance rates, with the highest acceptance in NSAID naïve patients (96.1%).</p><p><strong>Discussion: </strong>PGx NSAIDs CDS alert acceptance rates were higher compared to general CDS acceptance rates. This study highlights opportunities for continuous improvement including optimizing alert modality, modifying alert criteria to include look-back periods, and implementing genetically adapted ordersets.</p><p><strong>Conclusion: </strong>The initial acceptance rate of PGx NSAIDs CDS alerts was 62.6%, however, with significantly higher acceptance rates in NSAID naïve patients (62.6% vs 96.1%, <i>P</i> < .001). Integration of CDS is vital to the successful implementation of PGx in clinical practice.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf112"},"PeriodicalIF":3.4,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12492481/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145233287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Opportunities, barriers, and remedies for implementing REDCap integration with electronic health records via Fast Healthcare Interoperability Resources (FHIR). 通过快速医疗保健互操作性资源(FHIR)实现REDCap与电子健康记录集成的机会、障碍和补救措施。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-03 eCollection Date: 2025-10-01 DOI: 10.1093/jamiaopen/ooaf111
Alex C Cheng, Cathy Shyr, Adam Lewis, Francesco Delacqua, Teresa Bosler, Mary K Banasiewicz, Robert Taylor, Christopher J Lindsell, Paul A Harris

Objective: Accelerate adoption of clinical research technology that obtains electronic health record (EHR) data through HL7 Fast Healthcare Interoperability Resources (FHIR).

Materials and methods: Based on experience helping institutions implement REDCap-EHR integration and surveys of users and potential users, we discuss the technical and organizational barriers to adoption with strategies for remediation.

Results: With strong demand from researchers, the 21st Century Cures Act Final Rule in place, and REDCap software already in use at most research organizations, the environment seems ideal for REDCap-EHR integration for automated data exchange. However, concerns from information technology and regulatory leaders often slow progress and restrict how and when data from the EHR can be used.

Discussion and conclusion: While technological controls can help alleviate concerns about FHIR applications used in research, we have found that messaging, education, and extramural funding remain the strongest drivers of adoption.

目的:通过HL7快速医疗互操作性资源(FHIR)加速临床研究技术的采用,获取电子健康记录(EHR)数据。材料和方法:基于帮助机构实施REDCap-EHR整合的经验,以及对用户和潜在用户的调查,我们讨论了采用补救策略的技术和组织障碍。结果:由于研究人员的强烈需求,21世纪治愈法案的最终规则已经到位,REDCap软件已经在大多数研究机构中使用,REDCap- ehr集成自动化数据交换的环境似乎是理想的。然而,来自信息技术和监管领导人的担忧往往会减缓进展,并限制电子病历数据的使用方式和时间。讨论和结论:虽然技术控制可以帮助减轻对研究中使用的FHIR应用程序的担忧,但我们发现消息传递、教育和外部资金仍然是采用FHIR的最强大驱动因素。
{"title":"Opportunities, barriers, and remedies for implementing REDCap integration with electronic health records via Fast Healthcare Interoperability Resources (FHIR).","authors":"Alex C Cheng, Cathy Shyr, Adam Lewis, Francesco Delacqua, Teresa Bosler, Mary K Banasiewicz, Robert Taylor, Christopher J Lindsell, Paul A Harris","doi":"10.1093/jamiaopen/ooaf111","DOIUrl":"10.1093/jamiaopen/ooaf111","url":null,"abstract":"<p><strong>Objective: </strong>Accelerate adoption of clinical research technology that obtains electronic health record (EHR) data through HL7 Fast Healthcare Interoperability Resources (FHIR).</p><p><strong>Materials and methods: </strong>Based on experience helping institutions implement REDCap-EHR integration and surveys of users and potential users, we discuss the technical and organizational barriers to adoption with strategies for remediation.</p><p><strong>Results: </strong>With strong demand from researchers, the 21st Century Cures Act Final Rule in place, and REDCap software already in use at most research organizations, the environment seems ideal for REDCap-EHR integration for automated data exchange. However, concerns from information technology and regulatory leaders often slow progress and restrict how and when data from the EHR can be used.</p><p><strong>Discussion and conclusion: </strong>While technological controls can help alleviate concerns about FHIR applications used in research, we have found that messaging, education, and extramural funding remain the strongest drivers of adoption.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf111"},"PeriodicalIF":3.4,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12494410/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145233256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthetic data for pharmacogenetics: enabling scalable and secure research. 药物遗传学合成数据:实现可扩展和安全的研究。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-03 eCollection Date: 2025-10-01 DOI: 10.1093/jamiaopen/ooaf107
Marko Miletic, Anna Bollinger, Samuel S Allemann, Murat Sariyar

Objective: This study evaluates the performance of 7 synthetic data generation (SDG) methods-synthpop, avatar, copula, copulagan, ctgan, tvae, and the large language models-based tabula-for supporting pharmacogenetics (PGx) research.

Materials and methods: We used PGx profiles from 142 patients with adverse drug reactions or therapeutic failures, considering 2 scenarios: (1) a high-dimensional genotype dataset (104 variables) and (2) a phenotype dataset (24 variables). Models were assessed for (1) broad utility using propensity score mean squared error ( pMSE ), (2) specific utility via weighted F 1 score in a Train-Synthetic-Test-Real framework, and (3) privacy risk as ε-identifiability.

Results: Copula and synthpop consistently achieved strong performance across both datasets, combining low ε-identifiability (0.25-0.35) with competitive utility. Deep learning models like tabula and tvae trained for 10 000 epochs achieved lower pMSE but had higher ε-identifiability (>0.4) and limited gains in predictive performance. Specific utility was only weakly correlated with broad utility, indicating that distributional fidelity does not ensure predictive relevance. Copula and synthpop often outperformed original data in weighted F 1 scores, especially under noise or data imbalance.

Discussion: While deep learning models can achieve high distributional fidelity ( pMSE ), they often incur elevated ε-identifiability, raising privacy concerns. Traditional methods like copula and synthpop consistently offer robust utility and lower re-identification risk, particularly for high-dimensional data. Importantly, general utility does not predict specific utility ( F 1 score), emphasizing the need for multimetric evaluation.

Conclusion: No single SDG method dominated across all criteria. For privacy-sensitive PGx applications, classical methods such as copula and synthpop offer a reliable trade-off between utility and privacy, making them preferable for high-dimensional, limited-sample settings.

目的:评价7种合成数据生成(SDG)方法(synthpop、avatar、copula、copulagan、ctgan、tvae以及基于大语言模型的表格)在支持药物遗传学(PGx)研究中的性能。材料和方法:我们使用了142例药物不良反应或治疗失败患者的PGx谱,考虑了两种情况:(1)高维基因型数据集(104个变量)和(2)表型数据集(24个变量)。(1)使用倾向得分均方误差(pMSE)评估模型的广泛效用,(2)使用训练-合成-测试-真实框架中的加权f1得分评估模型的特定效用,以及(3)使用ε-可识别性评估模型的隐私风险。结果:Copula和synthpop在两个数据集上都取得了良好的表现,结合了低ε-可识别性(0.25-0.35)和竞争性效用。像tabula和tvae这样经过10000次训练的深度学习模型获得了较低的pMSE,但具有较高的ε-可识别性(>.4),并且预测性能的收益有限。具体效用仅与广泛效用弱相关,表明分布保真度不能确保预测相关性。Copula和synthpop在加权f1分数上往往优于原始数据,特别是在噪声或数据不平衡的情况下。讨论:虽然深度学习模型可以实现高分布保真度(pMSE),但它们通常会导致更高的ε-可识别性,从而引起隐私问题。copula和synthpop等传统方法始终提供强大的实用程序和较低的重新识别风险,特别是对于高维数据。重要的是,一般效用不能预测特定效用(f1分数),强调了多指标评估的必要性。结论:没有单一的SDG方法在所有标准中占主导地位。对于隐私敏感的PGx应用程序,copula和synthpop等经典方法在实用程序和隐私之间提供了可靠的权衡,使它们更适合高维、有限样本设置。
{"title":"Synthetic data for pharmacogenetics: enabling scalable and secure research.","authors":"Marko Miletic, Anna Bollinger, Samuel S Allemann, Murat Sariyar","doi":"10.1093/jamiaopen/ooaf107","DOIUrl":"10.1093/jamiaopen/ooaf107","url":null,"abstract":"<p><strong>Objective: </strong>This study evaluates the performance of 7 synthetic data generation (SDG) methods-synthpop, avatar, copula, copulagan, ctgan, tvae, and the large language models-based tabula-for supporting pharmacogenetics (PGx) research.</p><p><strong>Materials and methods: </strong>We used PGx profiles from 142 patients with adverse drug reactions or therapeutic failures, considering 2 scenarios: (1) a high-dimensional genotype dataset (104 variables) and (2) a phenotype dataset (24 variables). Models were assessed for (1) broad utility using propensity score mean squared error ( <math><mi>pMSE</mi></math> ), (2) specific utility via weighted <math> <mrow> <msub><mrow><mi>F</mi></mrow> <mrow><mn>1</mn></mrow> </msub> </mrow> </math> score in a Train-Synthetic-Test-Real framework, and (3) privacy risk as ε-identifiability.</p><p><strong>Results: </strong>Copula and synthpop consistently achieved strong performance across both datasets, combining low ε-identifiability (0.25-0.35) with competitive utility. Deep learning models like tabula and tvae trained for 10 000 epochs achieved lower <math><mi>pMSE</mi></math> but had higher ε-identifiability (>0.4) and limited gains in predictive performance. Specific utility was only weakly correlated with broad utility, indicating that distributional fidelity does not ensure predictive relevance. Copula and synthpop often outperformed original data in weighted <math> <mrow> <msub><mrow><mi>F</mi></mrow> <mrow><mn>1</mn></mrow> </msub> </mrow> </math> scores, especially under noise or data imbalance.</p><p><strong>Discussion: </strong>While deep learning models can achieve high distributional fidelity ( <math><mi>pMSE</mi></math> ), they often incur elevated ε-identifiability, raising privacy concerns. Traditional methods like copula and synthpop consistently offer robust utility and lower re-identification risk, particularly for high-dimensional data. Importantly, general utility does not predict specific utility ( <math> <mrow> <msub><mrow><mi>F</mi></mrow> <mrow><mn>1</mn></mrow> </msub> </mrow> </math> score), emphasizing the need for multimetric evaluation.</p><p><strong>Conclusion: </strong>No single SDG method dominated across all criteria. For privacy-sensitive PGx applications, classical methods such as copula and synthpop offer a reliable trade-off between utility and privacy, making them preferable for high-dimensional, limited-sample settings.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf107"},"PeriodicalIF":3.4,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12492482/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145233431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adding employment status in the medical record demonstrates its importance as a social determinant of health. 在医疗记录中增加就业状况表明其作为健康的社会决定因素的重要性。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-01 DOI: 10.1093/jamiaopen/ooaf108
Laura E Breeher, Samantha Westphal, Tammy Green, Alzhraa Abbas, Clayton T Cowl

Objective: To share findings of a quality improvement initiative to capture employment status in the electronic medical record (EMR) to mitigate the potential impact of work loss on the health of patients by utilizing the results to identify eligible individuals for resources to assist return-to-work efforts.

Materials and methods: Patients self-identified employment status through a structured new social determinants of health (SDOH) question within the EMR. An electronic outreach campaign was developed to provide information via the patient portal detailing services within healthcare and the community that could benefit the patients.

Results: Over the course of 12 months, 2059 patients were identified from the employment SDOH question. Resources to support stay at work and return to work efforts were provided to patients through an automated electronic portal campaign with 87% of patients reading the message and 7% engaging with a healthcare return to work case manager.

Discussion: Loss of employment has detrimental impacts on individual and population health. Most EMRs do not capture information on employment status. Adding this simple question identified individuals with potential gaps in SDOH, and in this case allowed specific resources to be shared with patients with an illness or injury that was acutely impacting work.

Conclusion: Medical center decision makers and EMR programmers should consider adding employment status as a social determinant of health.

目标:分享一项质量改进倡议的发现,该倡议旨在在电子病历(EMR)中记录就业状况,利用结果确定符合条件的个人,以获得资源,帮助他们重返工作岗位,从而减轻失业对患者健康的潜在影响。材料和方法:患者自我确定的就业状况,通过一个结构化的新的健康社会决定因素(SDOH)问题在电子病历。开展了一项电子外联运动,通过患者门户网站提供信息,详细说明医疗保健和社区内可以使患者受益的服务。结果:在12个月的时间里,从就业SDOH问题中确定了2059例患者。通过自动电子门户活动向患者提供了支持留在工作岗位和重返工作岗位的资源,87%的患者阅读了消息,7%的患者与医疗保健重返工作岗位案例管理人员进行了互动。讨论:失业对个人和人口健康有不利影响。大多数电子病历不包含就业状况信息。添加这个简单的问题可以确定在SDOH方面存在潜在差距的个体,并且在这种情况下,可以与患有严重影响工作的疾病或受伤的患者共享特定的资源。结论:医疗中心决策者和电子病历编制人员应考虑将就业状况作为健康的社会决定因素。
{"title":"Adding employment status in the medical record demonstrates its importance as a social determinant of health.","authors":"Laura E Breeher, Samantha Westphal, Tammy Green, Alzhraa Abbas, Clayton T Cowl","doi":"10.1093/jamiaopen/ooaf108","DOIUrl":"10.1093/jamiaopen/ooaf108","url":null,"abstract":"<p><strong>Objective: </strong>To share findings of a quality improvement initiative to capture employment status in the electronic medical record (EMR) to mitigate the potential impact of work loss on the health of patients by utilizing the results to identify eligible individuals for resources to assist return-to-work efforts.</p><p><strong>Materials and methods: </strong>Patients self-identified employment status through a structured new social determinants of health (SDOH) question within the EMR. An electronic outreach campaign was developed to provide information via the patient portal detailing services within healthcare and the community that could benefit the patients.</p><p><strong>Results: </strong>Over the course of 12 months, 2059 patients were identified from the employment SDOH question. Resources to support stay at work and return to work efforts were provided to patients through an automated electronic portal campaign with 87% of patients reading the message and 7% engaging with a healthcare return to work case manager.</p><p><strong>Discussion: </strong>Loss of employment has detrimental impacts on individual and population health. Most EMRs do not capture information on employment status. Adding this simple question identified individuals with potential gaps in SDOH, and in this case allowed specific resources to be shared with patients with an illness or injury that was acutely impacting work.</p><p><strong>Conclusion: </strong>Medical center decision makers and EMR programmers should consider adding employment status as a social determinant of health.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf108"},"PeriodicalIF":3.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12488229/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145213795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Workflows to automate covariate-adaptive randomization in REDCap via data entry triggers. 在REDCap中通过数据输入触发器实现协变量自适应随机化自动化的工作流程。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-01 DOI: 10.1093/jamiaopen/ooaf110
Jacob M Schauer, Marc O Broxton, Luke V Rasmussen, Gregory Swann, Michael E Newcomb, Jody D Ciolino

Objective: Covariate-adaptive randomization algorithms (CARAs) can reduce covariate imbalance in randomized controlled trials (RCTs), but a lack of integration into Research Electronic Data Capture (REDCap) has limited their use. We developed a software pipeline to seamlessly integrate CARAs into REDCap as part of the all2GETHER study, a 2-armed RCT concerning HIV prevention.

Materials and methods: Leveraging REDCap's Data Entry Trigger and a separate server, we implemented software in PHP and R to automate randomizations for all2GETHER. Randomizations were triggered by saving a specific REDCap form and were automatically communicated to unblinded study personnel.

Results: Study arms were highly comparable, with differences across covariates characterized by Cohen's d = 0.003 for continuous variables and risk differences <2.4% for categorical/binary variables.

Conclusions: Our pipeline proved effective at reducing covariate imbalance with minimal additional effort for study personnel.

Discussion: This pipeline is reproducible and could be used by other RCTs that collect data via REDCap.

目的:协变量自适应随机化算法(CARAs)可以减少随机对照试验(rct)中的协变量不平衡,但缺乏与研究电子数据捕获(REDCap)的集成限制了其使用。我们开发了一个软件管道,将CARAs无缝集成到REDCap中,作为all2together研究的一部分,这是一项关于艾滋病毒预防的双臂随机对照试验。材料和方法:利用REDCap的数据输入触发器和一个单独的服务器,我们在PHP和R中实现了all2together的自动随机化软件。随机化是通过保存特定的REDCap表格触发的,并自动传达给非盲研究人员。结果:研究组具有高度可比性,连续变量和风险差异的协变量差异以Cohen’s d = 0.003为特征。结论:我们的管道证明在减少协变量不平衡方面有效,研究人员的额外努力最少。讨论:这个管道是可重复的,可以被其他通过REDCap收集数据的随机对照试验使用。
{"title":"Workflows to automate covariate-adaptive randomization in REDCap via data entry triggers.","authors":"Jacob M Schauer, Marc O Broxton, Luke V Rasmussen, Gregory Swann, Michael E Newcomb, Jody D Ciolino","doi":"10.1093/jamiaopen/ooaf110","DOIUrl":"10.1093/jamiaopen/ooaf110","url":null,"abstract":"<p><strong>Objective: </strong>Covariate-adaptive randomization algorithms (CARAs) can reduce covariate imbalance in randomized controlled trials (RCTs), but a lack of integration into Research Electronic Data Capture (REDCap) has limited their use. We developed a software pipeline to seamlessly integrate CARAs into REDCap as part of the all2GETHER study, a 2-armed RCT concerning HIV prevention.</p><p><strong>Materials and methods: </strong>Leveraging REDCap's Data Entry Trigger and a separate server, we implemented software in PHP and R to automate randomizations for all2GETHER. Randomizations were triggered by saving a specific REDCap form and were automatically communicated to unblinded study personnel.</p><p><strong>Results: </strong>Study arms were highly comparable, with differences across covariates characterized by Cohen's <i>d</i> = 0.003 for continuous variables and risk differences <2.4% for categorical/binary variables.</p><p><strong>Conclusions: </strong>Our pipeline proved effective at reducing covariate imbalance with minimal additional effort for study personnel.</p><p><strong>Discussion: </strong>This pipeline is reproducible and could be used by other RCTs that collect data via REDCap.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf110"},"PeriodicalIF":3.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12486239/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145213994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging open-source large language models for clinical information extraction in resource-constrained settings. 利用开源大型语言模型在资源受限的环境中提取临床信息。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-01 DOI: 10.1093/jamiaopen/ooaf109
Luc Builtjes, Joeran Bosma, Mathias Prokop, Bram van Ginneken, Alessa Hering

Objective: We aimed to evaluate the zero-shot performance of open-source generative large language models (LLMs) on clinical information extraction from Dutch medical reports using the Diagnostic Report Analysis: General Optimization of NLP (DRAGON) benchmark.

Methods: We developed and released the llm_extractinator framework, a scalable, open-source tool for automating information extraction from clinical texts using LLMs. We evaluated 9 multilingual open-source LLMs across 28 tasks in the DRAGON benchmark, covering classification, regression, and named entity recognition (NER). All tasks were performed in a zero-shot setting. Model performance was quantified using task-specific metrics and aggregated into a DRAGON utility score. Additionally, we investigated the effect of in-context translation to English.

Results: Llama-3.3-70B achieved the highest utility score (0.760), followed by Phi-4-14B (0.751), Qwen-2.5-14B (0.748), and DeepSeek-R1-14B (0.744). These models outperformed or matched a fine-tuned RoBERTa baseline on 17 of 28 tasks, particularly in regression and structured classification. NER performance was consistently low across all models. Translation to English consistently reduced performance.

Discussion: Generative LLMs demonstrated strong zero-shot capabilities on clinical natural language processing tasks involving structured inference. Models around 14B parameters performed well overall, with Llama-3.3-70B leading but at high computational cost. Generative models excelled in regression tasks, but were hindered by token-level output formats for NER. Translation to English reduced performance, emphasizing the need for native language support.

Conclusion: Open-source generative LLMs provide a viable zero-shot alternative for clinical information extraction from Dutch medical texts, particularly in low-resource and multilingual settings.

目的:我们旨在使用诊断报告分析:NLP的一般优化(DRAGON)基准来评估开源生成大语言模型(LLMs)在从荷兰医学报告中提取临床信息方面的零射击性能。方法:我们开发并发布了llm_extractinator框架,这是一个可扩展的开源工具,用于使用llm从临床文本中自动提取信息。我们在DRAGON基准测试中评估了跨28个任务的9个多语言开源llm,涵盖分类、回归和命名实体识别(NER)。所有的任务都是在零射击的情况下完成的。使用特定于任务的度量对模型性能进行量化,并汇总到DRAGON效用评分中。此外,我们还研究了语境翻译对英语的影响。结果:Llama-3.3-70B的效用得分最高(0.760),其次是Phi-4-14B(0.751)、qwen2 .5- 14b(0.748)和DeepSeek-R1-14B(0.744)。这些模型在28项任务中的17项上优于或匹配微调的RoBERTa基线,特别是在回归和结构化分类方面。在所有模型中,NER的性能一直很低。翻译成英语会持续降低表现。讨论:生成法学硕士在涉及结构化推理的临床自然语言处理任务中表现出强大的零射击能力。14B参数附近的模型总体表现良好,其中羊驼-3.3- 70b领先,但计算成本较高。生成模型在回归任务中表现出色,但受到NER的令牌级输出格式的阻碍。翻译成英语降低了性能,强调需要母语支持。结论:开源生成法学硕士为从荷兰医学文本中提取临床信息提供了一个可行的零射击替代方案,特别是在低资源和多语言环境中。
{"title":"Leveraging open-source large language models for clinical information extraction in resource-constrained settings.","authors":"Luc Builtjes, Joeran Bosma, Mathias Prokop, Bram van Ginneken, Alessa Hering","doi":"10.1093/jamiaopen/ooaf109","DOIUrl":"10.1093/jamiaopen/ooaf109","url":null,"abstract":"<p><strong>Objective: </strong>We aimed to evaluate the zero-shot performance of open-source generative large language models (LLMs) on clinical information extraction from Dutch medical reports using the Diagnostic Report Analysis: General Optimization of NLP (DRAGON) benchmark.</p><p><strong>Methods: </strong>We developed and released the llm_extractinator framework, a scalable, open-source tool for automating information extraction from clinical texts using LLMs. We evaluated 9 multilingual open-source LLMs across 28 tasks in the DRAGON benchmark, covering classification, regression, and named entity recognition (NER). All tasks were performed in a zero-shot setting. Model performance was quantified using task-specific metrics and aggregated into a DRAGON utility score. Additionally, we investigated the effect of in-context translation to English.</p><p><strong>Results: </strong>Llama-3.3-70B achieved the highest utility score (0.760), followed by Phi-4-14B (0.751), Qwen-2.5-14B (0.748), and DeepSeek-R1-14B (0.744). These models outperformed or matched a fine-tuned RoBERTa baseline on 17 of 28 tasks, particularly in regression and structured classification. NER performance was consistently low across all models. Translation to English consistently reduced performance.</p><p><strong>Discussion: </strong>Generative LLMs demonstrated strong zero-shot capabilities on clinical natural language processing tasks involving structured inference. Models around 14B parameters performed well overall, with Llama-3.3-70B leading but at high computational cost. Generative models excelled in regression tasks, but were hindered by token-level output formats for NER. Translation to English reduced performance, emphasizing the need for native language support.</p><p><strong>Conclusion: </strong>Open-source generative LLMs provide a viable zero-shot alternative for clinical information extraction from Dutch medical texts, particularly in low-resource and multilingual settings.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf109"},"PeriodicalIF":3.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12488231/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145213796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Early auxiliary diagnosis model for chest pain triad based on artificial intelligence multimodal fusion. 基于人工智能多模态融合的胸痛三联征早期辅助诊断模型。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-01 DOI: 10.1093/jamiaopen/ooaf114
Jun Tang, Fang Chen, Dongdong Wu

Objectives: Acute chest pain is a common presentation in the emergency department, characterized by sudden onset with high morbidity and mortality. Traditional diagnostic methods, such as computed tomography (CT) and CT angiography (CTA), are often time-consuming and fail to meet the urgent need for rapid triage in emergency settings.

Materials and methods: We developed a multimodal model that integrates Bio-ClinicalBERT and ensemble learning (AdaBoost, Gradient boosting, and XGBoost) based on 41 382 patient data from April 1, 2013 to April 1, 2025 at Chongqing Daping Hospital. By integrating clinical texts and laboratory indicators, the model aims to classify the 3 major causes of fatal chest pain (acute coronary syndrome, pulmonary embolism, and aortic dissection), as well as other causes of chest pain, aiding rapid triage. We adopt strict data preprocessing and rank importance feature selection.

Results: The multimodal fusion model based on Gradient boosting exhibits the best performance: accuracy of 88.40%, area under the curve of 0.951, F1-score of 74.56%, precision of 77.50%, and recall of 72.52%. SHapley Additive exPlanations (SHAP) analysis confirmed the clinical relevance of key features such as d-dimer and high-sensitivity troponin. When reducing the number of numerical features to 30 key indicators, the model enhanced robustness without compromising performance.

Discussion and conclusion: We developed an artificial intelligence model for chest pain classification that effectively addresses the problem of overlapping clinical symptoms through multimodal fusion, and the model has high accuracy. However, future work needs to better integrate model development with clinical workflows and practical constraints.

目的:急性胸痛是急诊科常见的症状,其特点是发病突然,发病率和死亡率高。传统的诊断方法,如计算机断层扫描(CT)和CT血管造影(CTA),往往耗时,不能满足紧急情况下快速分诊的迫切需要。材料和方法:基于重庆大平医院2013年4月1日至2025年4月1日的41 382例患者数据,我们开发了一个集成了Bio-ClinicalBERT和集成学习(AdaBoost, Gradient boosting和XGBoost)的多模式模型。通过整合临床文献和实验室指标,该模型旨在对致死性胸痛的3个主要原因(急性冠状动脉综合征、肺栓塞和主动脉夹层)以及其他胸痛原因进行分类,帮助快速分诊。我们采用严格的数据预处理和排序重要特征选择。结果:基于梯度增强的多模态融合模型的准确率为88.40%,曲线下面积为0.951,f1评分为74.56%,准确率为77.50%,召回率为72.52%。SHapley加性解释(SHAP)分析证实了d-二聚体和高敏感性肌钙蛋白等关键特征的临床相关性。当将数字特征的数量减少到30个关键指标时,该模型在不影响性能的情况下增强了鲁棒性。讨论与结论:我们开发了一个胸痛分类的人工智能模型,通过多模态融合有效解决了临床症状重叠的问题,模型具有较高的准确率。然而,未来的工作需要更好地将模型开发与临床工作流程和实际限制相结合。
{"title":"Early auxiliary diagnosis model for chest pain triad based on artificial intelligence multimodal fusion.","authors":"Jun Tang, Fang Chen, Dongdong Wu","doi":"10.1093/jamiaopen/ooaf114","DOIUrl":"10.1093/jamiaopen/ooaf114","url":null,"abstract":"<p><strong>Objectives: </strong>Acute chest pain is a common presentation in the emergency department, characterized by sudden onset with high morbidity and mortality. Traditional diagnostic methods, such as computed tomography (CT) and CT angiography (CTA), are often time-consuming and fail to meet the urgent need for rapid triage in emergency settings.</p><p><strong>Materials and methods: </strong>We developed a multimodal model that integrates Bio-ClinicalBERT and ensemble learning (AdaBoost, Gradient boosting, and XGBoost) based on 41 382 patient data from April 1, 2013 to April 1, 2025 at Chongqing Daping Hospital. By integrating clinical texts and laboratory indicators, the model aims to classify the 3 major causes of fatal chest pain (acute coronary syndrome, pulmonary embolism, and aortic dissection), as well as other causes of chest pain, aiding rapid triage. We adopt strict data preprocessing and rank importance feature selection.</p><p><strong>Results: </strong>The multimodal fusion model based on Gradient boosting exhibits the best performance: accuracy of 88.40%, area under the curve of 0.951, F1-score of 74.56%, precision of 77.50%, and recall of 72.52%. SHapley Additive exPlanations (SHAP) analysis confirmed the clinical relevance of key features such as d-dimer and high-sensitivity troponin. When reducing the number of numerical features to 30 key indicators, the model enhanced robustness without compromising performance.</p><p><strong>Discussion and conclusion: </strong>We developed an artificial intelligence model for chest pain classification that effectively addresses the problem of overlapping clinical symptoms through multimodal fusion, and the model has high accuracy. However, future work needs to better integrate model development with clinical workflows and practical constraints.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf114"},"PeriodicalIF":3.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12486235/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145213844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated phenotyping of congenital heart disease for dynamic patient aggregation and outcome reporting. 用于动态患者聚集和结果报告的先天性心脏病自动表型分析。
IF 3.4 Q2 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-10-01 DOI: 10.1093/jamiaopen/ooaf106
Shuhei Toba, Taylor M Smith, Francesca Sperotto, Chrystalle Katte Carreon, Kwannapas Saengsin, Samuel Casella, Marlon Delgado, Peng Zeng, Stephen P Sanders, Audrey Dionne, Eric N Feins, Steven D Colan, John E Mayer, John N Kheir

Objectives: Accurate characterization of patients with congenital heart disease is fundamental to research, outcomes reporting, quality improvement, and clinical decision-making. Here we present an approach to computing the anatomy of patients with congenital heart disease based on the whole of their diagnostic and surgical codes.

Materials and methods: All diagnostic and procedure codes for patients cared for between 1981 and 2020 at Boston Children's Hospital were extracted from a database containing diagnostic codes from echocardiograms, and procedural codes from surgical and catheterization procedures. The pipeline sequentially (1) mapped each of the 7500 native codes to algorithm codes; (2) computed the parent anatomy for each study using a pre-defined hierarchy; (3) computed the parent anatomy for the patient, based on highest ranking parent anatomy; and (4) computed the subcategories and mandatory co-variate findings for each patient. Thereafter, diagnostic accuracy of 500 unseen patients was adjudicated against clinical documentation by clinical experts.

Results: A total of 514 541 echocardiograms on 161 735 patients were available for this study. Phenotypes of congenital cardiac diseases were assigned in 84 285 patients (52%), and the remainder were computed to have normal anatomy. Clinicians agreed with algorithm assignments in 96.4% (482 of 500 patients), with disagreements most often representing definitional differences. An interactive dashboard enabled by the output of this algorithm is presented.

Conclusions: The computation of detailed congenital heart defect phenotypes from raw diagnostic and procedure codes is possible with a high degree of accuracy and efficiency. This framework may enable tools to support interactive outcomes reporting and clinical decision support.

目的:先天性心脏病患者的准确特征是研究、结果报告、质量改进和临床决策的基础。在这里,我们提出了一种方法来计算先天性心脏病患者的解剖基于他们的诊断和手术代码。材料和方法:从包含超声心动图诊断代码以及外科和导管手术程序代码的数据库中提取1981年至2020年在波士顿儿童医院接受治疗的患者的所有诊断和程序代码。所述管道依次(1)将所述7500个本机代码中的每一个映射为算法代码;(2)使用预定义的层次结构计算每个研究的父结构;(3)根据最高亲本解剖结构计算患者亲本解剖结构;(4)计算每个患者的亚类和强制性协变量结果。此后,500名未见患者的诊断准确性由临床专家根据临床文件进行裁决。结果:16735例患者共获得514541张超声心动图。对84 285例(52%)先天性心脏病患者进行了表型分析,其余患者的解剖结构正常。临床医生同意算法分配的比例为96.4%(500名患者中的482名),而分歧通常代表定义上的差异。提出了一种基于该算法输出的交互式仪表板。结论:从原始诊断和程序代码中计算出详细的先天性心脏缺陷表型是可能的,具有高度的准确性和效率。该框架可以使工具支持交互式结果报告和临床决策支持。
{"title":"Automated phenotyping of congenital heart disease for dynamic patient aggregation and outcome reporting.","authors":"Shuhei Toba, Taylor M Smith, Francesca Sperotto, Chrystalle Katte Carreon, Kwannapas Saengsin, Samuel Casella, Marlon Delgado, Peng Zeng, Stephen P Sanders, Audrey Dionne, Eric N Feins, Steven D Colan, John E Mayer, John N Kheir","doi":"10.1093/jamiaopen/ooaf106","DOIUrl":"10.1093/jamiaopen/ooaf106","url":null,"abstract":"<p><strong>Objectives: </strong>Accurate characterization of patients with congenital heart disease is fundamental to research, outcomes reporting, quality improvement, and clinical decision-making. Here we present an approach to computing the anatomy of patients with congenital heart disease based on the whole of their diagnostic and surgical codes.</p><p><strong>Materials and methods: </strong>All diagnostic and procedure codes for patients cared for between 1981 and 2020 at Boston Children's Hospital were extracted from a database containing diagnostic codes from echocardiograms, and procedural codes from surgical and catheterization procedures. The pipeline sequentially (1) mapped each of the 7500 native codes to algorithm codes; (2) computed the parent anatomy for each study using a pre-defined hierarchy; (3) computed the parent anatomy for the patient, based on highest ranking parent anatomy; and (4) computed the subcategories and mandatory co-variate findings for each patient. Thereafter, diagnostic accuracy of 500 unseen patients was adjudicated against clinical documentation by clinical experts.</p><p><strong>Results: </strong>A total of 514 541 echocardiograms on 161 735 patients were available for this study. Phenotypes of congenital cardiac diseases were assigned in 84 285 patients (52%), and the remainder were computed to have normal anatomy. Clinicians agreed with algorithm assignments in 96.4% (482 of 500 patients), with disagreements most often representing definitional differences. An interactive dashboard enabled by the output of this algorithm is presented.</p><p><strong>Conclusions: </strong>The computation of detailed congenital heart defect phenotypes from raw diagnostic and procedure codes is possible with a high degree of accuracy and efficiency. This framework may enable tools to support interactive outcomes reporting and clinical decision support.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"8 5","pages":"ooaf106"},"PeriodicalIF":3.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12486236/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145213774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
JAMIA Open
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1