Sky Corby, Joan S Ash, Rebecca M Jungbauer, Gretchen Scholl, Sarah Florig, Vishnu Mohan, Jeffrey A Gold
Background: Electronic health records (EHRs) can aid in provider efficiency, but may also lead to unintended consequences, such as documentation burden and increased length of notes. To combat issues related to documentation, copying and pasting (CP) and copying or carrying forward (CF) are tools that have been used to aid in documentation burden. Multiple studies have identified the benefits and challenges of using these tools; however, few studies have identified the unintended consequences of CP and CF, and how the adoption of these tools may affect users.
Objective: The objective was to describe providers' perceptions and use of copying tools available in the EHR and describe their suggestions for improvement on these copying tools.
Methods: Research team members conducted semistructured interviews with faculty members, advanced practice providers, residents or fellow trainees, and medical students at a single academic health sciences center. The Diffusion of Innovations Theory of Unintended Consequences guided the analysis and interpretation of interview results.
Results: A total of 22 semistructured interviews were conducted in 2023 and analyzed during 2024. The findings showed that respondents use and value these tools for efficiency and communication purposes. The negative unintended consequences include inaccuracies and errors in documentation and increased patient safety risks. Some respondents experience inner angst or moral injury related to using CP/CF, but they feel that they must use them to satisfy organizational requirements surrounding documentation. The respondents suggested that artificial intelligence will likely help improve documentation tools, as would further training around these types of documentation tools.
Conclusions: Some respondents noted feeling both internal and external pressures that influenced when and how they use CP/CF. Respondents noted that they value EHR copying tools for efficiency purposes, but they also understand the risks involved. This tension may lead to moral angst or moral injury. They offered numerous suggestions for lowering the risk, especially by improving the documentation capabilities of the EHR through artificial intelligence. Future research should investigate both technical and educational solutions to relieve the documentation burden and moral angst they are experiencing.
{"title":"Copy Tools in the Electronic Health Record: Perceptions, Implications, and Future Directions.","authors":"Sky Corby, Joan S Ash, Rebecca M Jungbauer, Gretchen Scholl, Sarah Florig, Vishnu Mohan, Jeffrey A Gold","doi":"10.2196/78502","DOIUrl":"10.2196/78502","url":null,"abstract":"<p><strong>Background: </strong>Electronic health records (EHRs) can aid in provider efficiency, but may also lead to unintended consequences, such as documentation burden and increased length of notes. To combat issues related to documentation, copying and pasting (CP) and copying or carrying forward (CF) are tools that have been used to aid in documentation burden. Multiple studies have identified the benefits and challenges of using these tools; however, few studies have identified the unintended consequences of CP and CF, and how the adoption of these tools may affect users.</p><p><strong>Objective: </strong>The objective was to describe providers' perceptions and use of copying tools available in the EHR and describe their suggestions for improvement on these copying tools.</p><p><strong>Methods: </strong>Research team members conducted semistructured interviews with faculty members, advanced practice providers, residents or fellow trainees, and medical students at a single academic health sciences center. The Diffusion of Innovations Theory of Unintended Consequences guided the analysis and interpretation of interview results.</p><p><strong>Results: </strong>A total of 22 semistructured interviews were conducted in 2023 and analyzed during 2024. The findings showed that respondents use and value these tools for efficiency and communication purposes. The negative unintended consequences include inaccuracies and errors in documentation and increased patient safety risks. Some respondents experience inner angst or moral injury related to using CP/CF, but they feel that they must use them to satisfy organizational requirements surrounding documentation. The respondents suggested that artificial intelligence will likely help improve documentation tools, as would further training around these types of documentation tools.</p><p><strong>Conclusions: </strong>Some respondents noted feeling both internal and external pressures that influenced when and how they use CP/CF. Respondents noted that they value EHR copying tools for efficiency purposes, but they also understand the risks involved. This tension may lead to moral angst or moral injury. They offered numerous suggestions for lowering the risk, especially by improving the documentation capabilities of the EHR through artificial intelligence. Future research should investigate both technical and educational solutions to relieve the documentation burden and moral angst they are experiencing.</p>","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e78502"},"PeriodicalIF":3.8,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12759298/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145795363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background: Accurately predicting the survival outcomes of patients with lung cancer receiving chemotherapy remains challenging.
Objective: To improve clinical management of this population, this study developed a multivariate machine learning (ML) model to assess all-cause mortality risk in chemotherapy-treated patients with lung cancer.
Methods: This study retrospectively recruited 1278 postchemotherapy patients with lung cancer from Guangzhou Chest Hospital between 2017 and 2019. Candidate features such as demographic characteristics, environmental exposures, clinical information, and patient-reported symptoms were collected via questionnaires and the electronic medical record system. The survival status and the deceased date were investigated twice a year. A total of 84 predictive models were constructed on the training set using 5 ML algorithms either individually or in pairwise combinations. The concordance index was used to identify the optimal model on the testing set, with performance validated via receiver operating characteristic curves, calibration curves, and decision curve analysis. Additionally, Shapley Additive Explanations and restricted cubic splines were applied for feature attribution analysis.
Results: The optimal model ultimately retained 21 prognosis-association features, including age, sex, BMI, smoking status, environmental smoke, the MD Anderson Symptom Inventory for Lung Cancer total score trajectories, cluster of differentiation 56, TNM stage, histology, and prechemotherapy blood biomarkers. On the testing set, the model acquired a concordance index of 0.702 (95% CI 0.652-0.753). The decision curves demonstrated positive clinical benefit when the risk thresholds were 0.40-0.69, 0.62-0.99, and 0.72-0.99 for 1-, 3-, and 5-year mortality predictions, respectively. The calibration curves showed that the predicted mortality probabilities fluctuated around the observed probabilities, and the Brier scores for 1-, 3-, and 5-year predictions were 0.20, 0.18, and 0.11, respectively. The area under the curve of the model was 0.740, 0.777, and 0.915 for 1-, 3-, and 5-year mortality predictions, respectively. Interpretability feature attribution analysis revealed that the significant features could predict all-cause mortality risk in chemotherapy-treated patients with lung cancer.
Conclusions: Our ML models exhibited acceptable discrimination, calibration, and clinical benefit in predicting the mortality risk of chemotherapy-treated patients with lung cancer, which could help clinicians in personalized prognostic management.
背景:准确预测接受化疗的肺癌患者的生存结果仍然具有挑战性。目的:为了改善这一人群的临床管理,本研究建立了一个多变量机器学习(ML)模型来评估化疗肺癌患者的全因死亡率风险。方法:本研究回顾性招募2017 - 2019年广州胸科医院肺癌化疗后患者1278例。候选特征,如人口统计学特征、环境暴露、临床信息和患者报告的症状,通过问卷调查和电子病历系统收集。每年调查两次患者的生存状况和死亡日期。在训练集上,使用5种ML算法单独或成对组合构建了84个预测模型。使用一致性指数在测试集上识别最优模型,并通过受试者工作特征曲线、校准曲线和决策曲线分析验证其性能。此外,应用Shapley加性解释和受限三次样条进行特征归因分析。结果:最佳模型最终保留了21个预后相关特征,包括年龄、性别、BMI、吸烟状况、环境烟雾、MD安德森肺癌症状量表总分轨迹、分化聚类56、TNM分期、组织学和化疗前血液生物标志物。在检验集上,模型的一致性指数为0.702 (95% CI 0.652-0.753)。当1年、3年和5年死亡率预测的风险阈值分别为0.40-0.69、0.62-0.99和0.72-0.99时,决策曲线显示出积极的临床获益。校正曲线显示,预测的死亡率概率在观测概率周围波动,1年、3年和5年预测的Brier评分分别为0.20、0.18和0.11。1年、3年和5年死亡率预测的曲线下面积分别为0.740、0.777和0.915。可解释性特征归因分析显示,显著性特征可以预测化疗肺癌患者的全因死亡风险。结论:我们的ML模型在预测化疗肺癌患者死亡风险方面表现出可接受的区分、校准和临床益处,可以帮助临床医生进行个性化预后管理。
{"title":"A Machine Learning Approach to Predicting Mortality Risk in Chemotherapy-Treated Lung Cancer: Machine Learning Model Development and Validation.","authors":"Jianjun Zou, Jinyi Huang, Katie Lu, Ao Lin, Chen Xie, Jinrong Zhang, Boqi Rao, Zhi Li, Dongming Xie, Ling Lu, Feng Luo, Jinbin Chen, Lei Yang, Fuman Qiu, Xin Zhang, Yibin Deng, Jiachun Lu","doi":"10.2196/72424","DOIUrl":"10.2196/72424","url":null,"abstract":"<p><strong>Background: </strong>Accurately predicting the survival outcomes of patients with lung cancer receiving chemotherapy remains challenging.</p><p><strong>Objective: </strong>To improve clinical management of this population, this study developed a multivariate machine learning (ML) model to assess all-cause mortality risk in chemotherapy-treated patients with lung cancer.</p><p><strong>Methods: </strong>This study retrospectively recruited 1278 postchemotherapy patients with lung cancer from Guangzhou Chest Hospital between 2017 and 2019. Candidate features such as demographic characteristics, environmental exposures, clinical information, and patient-reported symptoms were collected via questionnaires and the electronic medical record system. The survival status and the deceased date were investigated twice a year. A total of 84 predictive models were constructed on the training set using 5 ML algorithms either individually or in pairwise combinations. The concordance index was used to identify the optimal model on the testing set, with performance validated via receiver operating characteristic curves, calibration curves, and decision curve analysis. Additionally, Shapley Additive Explanations and restricted cubic splines were applied for feature attribution analysis.</p><p><strong>Results: </strong>The optimal model ultimately retained 21 prognosis-association features, including age, sex, BMI, smoking status, environmental smoke, the MD Anderson Symptom Inventory for Lung Cancer total score trajectories, cluster of differentiation 56, TNM stage, histology, and prechemotherapy blood biomarkers. On the testing set, the model acquired a concordance index of 0.702 (95% CI 0.652-0.753). The decision curves demonstrated positive clinical benefit when the risk thresholds were 0.40-0.69, 0.62-0.99, and 0.72-0.99 for 1-, 3-, and 5-year mortality predictions, respectively. The calibration curves showed that the predicted mortality probabilities fluctuated around the observed probabilities, and the Brier scores for 1-, 3-, and 5-year predictions were 0.20, 0.18, and 0.11, respectively. The area under the curve of the model was 0.740, 0.777, and 0.915 for 1-, 3-, and 5-year mortality predictions, respectively. Interpretability feature attribution analysis revealed that the significant features could predict all-cause mortality risk in chemotherapy-treated patients with lung cancer.</p><p><strong>Conclusions: </strong>Our ML models exhibited acceptable discrimination, calibration, and clinical benefit in predicting the mortality risk of chemotherapy-treated patients with lung cancer, which could help clinicians in personalized prognostic management.</p>","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e72424"},"PeriodicalIF":3.8,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12757710/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145776708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background: Knee cartilage injury (KCI) poses significant challenges in the early clinical diagnosis process, primarily due to its high incidence, the complexity of healing, and the limited sensitivity of initial imaging modalities.
Objective: This study aims to employ magnetic resonance imaging and machine learning methods to enhance the classification accuracy of the classifier for KCI, improve the existing network structure, and demonstrate important clinical application value.
Methods: The proposed methodology is a multidimensional feature cross-level fusion classification network driven by the large separable kernel attention, which enables high-precision hierarchical diagnosis of KCI through deep learning. The network first fuses shallow high-resolution features with deep semantic features via the cross-level fusion module. Then, the large separable kernel attention module is embedded in the YOLOv8 network. This network utilizes the combined optimization of depth-separable and point-by-point convolutions to enhance features at multiple scales, thereby dramatically improving the hierarchical characterization of cartilage damage. Finally, five classifications of knee cartilage injuries are performed by classifiers.
Results: To overcome the limitations of network models trained with single-plane images, this study presents the first hospital-based multidimensional magnetic resonance imaging real dataset for KCI, on which the classification accuracy is 99.7%, the Kappa statistic is 99.6%, the F-measure is 99.7%, the sensitivity is 99.7%, and the specificity is 99.9%. The experimental results validate the feasibility of the proposed method.
Conclusions: The experimental outcomes confirm that the proposed methodology not only achieves exceptional performance in classifying knee cartilage injuries but also offers substantial improvements over existing techniques. This underscores its potential for clinical deployment in enhancing diagnostic precision and efficiency.
{"title":"Large Separable Kernel Attention-Driven Multidimensional Feature Cross-Level Fusion Classification Network of Knee Cartilage Injury: Algorithm Development and Validation.","authors":"Lirong Zhang, Hang Yu, Yating Yang","doi":"10.2196/79748","DOIUrl":"10.2196/79748","url":null,"abstract":"<p><strong>Background: </strong>Knee cartilage injury (KCI) poses significant challenges in the early clinical diagnosis process, primarily due to its high incidence, the complexity of healing, and the limited sensitivity of initial imaging modalities.</p><p><strong>Objective: </strong>This study aims to employ magnetic resonance imaging and machine learning methods to enhance the classification accuracy of the classifier for KCI, improve the existing network structure, and demonstrate important clinical application value.</p><p><strong>Methods: </strong>The proposed methodology is a multidimensional feature cross-level fusion classification network driven by the large separable kernel attention, which enables high-precision hierarchical diagnosis of KCI through deep learning. The network first fuses shallow high-resolution features with deep semantic features via the cross-level fusion module. Then, the large separable kernel attention module is embedded in the YOLOv8 network. This network utilizes the combined optimization of depth-separable and point-by-point convolutions to enhance features at multiple scales, thereby dramatically improving the hierarchical characterization of cartilage damage. Finally, five classifications of knee cartilage injuries are performed by classifiers.</p><p><strong>Results: </strong>To overcome the limitations of network models trained with single-plane images, this study presents the first hospital-based multidimensional magnetic resonance imaging real dataset for KCI, on which the classification accuracy is 99.7%, the Kappa statistic is 99.6%, the F-measure is 99.7%, the sensitivity is 99.7%, and the specificity is 99.9%. The experimental results validate the feasibility of the proposed method.</p><p><strong>Conclusions: </strong>The experimental outcomes confirm that the proposed methodology not only achieves exceptional performance in classifying knee cartilage injuries but also offers substantial improvements over existing techniques. This underscores its potential for clinical deployment in enhancing diagnostic precision and efficiency.</p>","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e79748"},"PeriodicalIF":3.8,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12711134/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145776752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Freddie Seba, Miriam Isola, Laura Mills, Mohan Zalake, Jacob Krive
Background: We designed learning assignments for students to develop knowledge, skills, and professional attitudes about generative artificial intelligence (AI) in 2 different Master's level courses in health informatics. Our innovative approach assumed that the students had no technical background or experience in using generative AI tools.
Objective: This study aims to offer generalizable methods and experiences on integration and assessment of generative AI content into the higher education's health informatics curricula. The study's central driver is the preparation of graduate students with generative AI tools, skills, ethical discernment, and critical thinking capacities aligned with the rapidly shifting job-market requirements, independent of graduate students' backgrounds and technical expertise.
Methods: During the semester, students completed a pretest and posttest to assess knowledge about generative AI. Reflections explored their expectations and experiences using generative AI to complete their assignments and projects during the semester. Strong emphasis was placed on building skills and professional attitudes by using generative AI. Student engagement in behavioral, emotional, and cognitive domains was explored via detailed analysis of student reflections by faculty.
Results: Students at the University of Illinois Chicago increased their knowledge about generative AI from 81% to 93% through research of the basic generative AI concepts, as evidenced from outcomes of the open-book pre-and posttests given at the beginning and end of the capstone course. University of San Francisco students also improved from 77% to 80% by the end of the semester. Faculty analysis of student reflections upon completion of the course revealed primary interests in the essentials of generative AI, AI transformations to information and knowledge, and organizational changes influenced by AI adoption in the health care organizations, with ethics being a primary driver of students' interests and engagement.
Conclusions: Data from student reflections provided insight into generative AI skills that students developed and that health informatics programs can consider incorporating into their curricula. Building competencies in generative AI will prepare students for the 21st century workforce and enable them to build skills employers are seeking in the new digital health environment.
背景:我们设计了学习作业,让学生在两个不同的健康信息学硕士课程中发展关于生成式人工智能(AI)的知识、技能和专业态度。我们的创新方法假设学生没有使用生成式人工智能工具的技术背景或经验。目的:本研究旨在为高等教育健康信息学课程整合与评估生成式人工智能内容提供可推广的方法和经验。该研究的核心驱动力是培养具有生成式人工智能工具、技能、道德洞察力和批判性思维能力的研究生,使其与快速变化的就业市场需求保持一致,独立于研究生的背景和技术专长。方法:在学期中,学生通过前测和后测来评估生成式人工智能的知识。反思探讨了他们在学期中使用生成式人工智能完成作业和项目的期望和经验。重点是通过使用生成式人工智能来培养技能和专业态度。通过教师对学生反映的详细分析,探讨了学生在行为、情感和认知领域的参与情况。结果:通过对生成式人工智能基本概念的研究,伊利诺伊大学芝加哥分校的学生对生成式人工智能的了解从81%增加到93%,这一点从顶点课程开始和结束时进行的开卷前测和后测的结果可以看出。到学期末,旧金山大学(University of San Francisco)的学生也从77%提高到了80%。教师对学生完成课程后反思的分析显示,他们对生成式人工智能的基本要素、人工智能向信息和知识的转化以及卫生保健组织中人工智能采用所影响的组织变革感兴趣,而道德是学生兴趣和参与的主要驱动力。结论:来自学生反思的数据提供了对学生发展的生成式人工智能技能的见解,健康信息学专业可以考虑将其纳入课程。在生成式人工智能方面培养能力将使学生为21世纪的劳动力做好准备,并使他们能够掌握雇主在新的数字健康环境中所寻求的技能。
{"title":"Incorporating Generative AI Into a Health Informatics Curriculum to Build 21st Century Competencies: Multisite Pre-Post Study.","authors":"Freddie Seba, Miriam Isola, Laura Mills, Mohan Zalake, Jacob Krive","doi":"10.2196/76507","DOIUrl":"10.2196/76507","url":null,"abstract":"<p><strong>Background: </strong>We designed learning assignments for students to develop knowledge, skills, and professional attitudes about generative artificial intelligence (AI) in 2 different Master's level courses in health informatics. Our innovative approach assumed that the students had no technical background or experience in using generative AI tools.</p><p><strong>Objective: </strong>This study aims to offer generalizable methods and experiences on integration and assessment of generative AI content into the higher education's health informatics curricula. The study's central driver is the preparation of graduate students with generative AI tools, skills, ethical discernment, and critical thinking capacities aligned with the rapidly shifting job-market requirements, independent of graduate students' backgrounds and technical expertise.</p><p><strong>Methods: </strong>During the semester, students completed a pretest and posttest to assess knowledge about generative AI. Reflections explored their expectations and experiences using generative AI to complete their assignments and projects during the semester. Strong emphasis was placed on building skills and professional attitudes by using generative AI. Student engagement in behavioral, emotional, and cognitive domains was explored via detailed analysis of student reflections by faculty.</p><p><strong>Results: </strong>Students at the University of Illinois Chicago increased their knowledge about generative AI from 81% to 93% through research of the basic generative AI concepts, as evidenced from outcomes of the open-book pre-and posttests given at the beginning and end of the capstone course. University of San Francisco students also improved from 77% to 80% by the end of the semester. Faculty analysis of student reflections upon completion of the course revealed primary interests in the essentials of generative AI, AI transformations to information and knowledge, and organizational changes influenced by AI adoption in the health care organizations, with ethics being a primary driver of students' interests and engagement.</p><p><strong>Conclusions: </strong>Data from student reflections provided insight into generative AI skills that students developed and that health informatics programs can consider incorporating into their curricula. Building competencies in generative AI will prepare students for the 21st century workforce and enable them to build skills employers are seeking in the new digital health environment.</p>","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e76507"},"PeriodicalIF":3.8,"publicationDate":"2025-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12707436/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145769877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sizhe Jasmine Chen, Da Xu, Derek K Hu, Paul Jen-Hwa Hu, Ting-Shuo Huang
Background: Metabolic dysfunction-associated fatty liver disease (MAFLD) is a leading cause of chronic disease and can progress to liver fibrosis or hepatocellular carcinoma. Its subtypes-obese, diabetic, and lean-are associated with varying degrees of fibrotic burden and different complications, yet the existing analytics methods often overlook its multisystem nature, intraphenotype variability, and disease dynamics. These limitations hinder accurate risk stratification and restrict personalized intervention planning.
Objective: This study developed a novel, 2-stage, contrastive learning-based method to predict the phenotype of MAFLD among adults. This method leverages multiview contrastive learning; it models individual heterogeneities and important relationships in clinical and survey-based data to predict phenotypes among adults, thus supporting clinical decision-making and personalized care.
Methods: Demographic, clinical, lifestyle, and genetic family history data of 4408 adults revealed how capturing essential relationships in patient data from different sources can transform individual-level representations into multiple, complementary views. Evaluation of the predictive efficacy of the proposed method in comparison with 8 prevalent methods relied on recall, precision, F1-score, and area under the curve values. Moreover, a Shapley additive explanation analysis was performed for interpretability.
Results: The proposed method consistently and significantly outperformed all benchmark methods. It attained the highest F1-score, showing a 32.8% improvement for nondiabetic MAFLD (0.531 vs 0.400) and 30.4% improvement for diabetic MAFLD (0.519 vs 0.398) over the respective best-performing benchmark. The results underscore the clinical value and utility of integrating clinical and survey-based data in the prediction of MAFLD phenotypes among adults.
Conclusions: The proposed method is a viable approach for MAFLD phenotype prediction. It is more effective in identifying at-risk adults than many prevalent data-driven analytics methods and thereby can enhance clinical decision-making and support patient-centric care and management.
背景:代谢功能障碍相关脂肪性肝病(MAFLD)是慢性疾病的主要原因,可发展为肝纤维化或肝细胞癌。其亚型(肥胖、糖尿病和消瘦)与不同程度的纤维化负担和不同的并发症相关,但现有的分析方法往往忽略了其多系统性质、表型内变异性和疾病动力学。这些限制阻碍了准确的风险分层,并限制了个性化的干预计划。目的:本研究开发了一种新的、两阶段的、基于对比学习的方法来预测成人MAFLD的表型。这种方法利用了多视角对比学习;它模拟了临床和基于调查的数据中的个体异质性和重要关系,以预测成人的表型,从而支持临床决策和个性化护理。方法:4408名成年人的人口统计、临床、生活方式和遗传家族史数据揭示了如何从不同来源的患者数据中捕捉基本关系,将个人层面的表征转化为多个互补的观点。与8种流行方法相比,该方法的预测效果评估依赖于召回率、精度、f1评分和曲线下面积值。此外,对可解释性进行了Shapley加性解释分析。结果:所提方法一致且显著优于所有基准方法。它获得了最高的f1评分,与各自的最佳表现基准相比,非糖尿病性MAFLD改善了32.8% (0.531 vs 0.400),糖尿病性MAFLD改善了30.4% (0.519 vs 0.398)。研究结果强调了综合临床和基于调查的数据预测成人MAFLD表型的临床价值和效用。结论:该方法是一种预测MAFLD表型的可行方法。它比许多流行的数据驱动分析方法更有效地识别有风险的成年人,从而可以增强临床决策并支持以患者为中心的护理和管理。
{"title":"Predicting Metabolic Dysfunction-Associated Fatty Liver Disease Phenotypes Among Adults: 2-Stage Contrastive Learning Method.","authors":"Sizhe Jasmine Chen, Da Xu, Derek K Hu, Paul Jen-Hwa Hu, Ting-Shuo Huang","doi":"10.2196/75747","DOIUrl":"10.2196/75747","url":null,"abstract":"<p><strong>Background: </strong>Metabolic dysfunction-associated fatty liver disease (MAFLD) is a leading cause of chronic disease and can progress to liver fibrosis or hepatocellular carcinoma. Its subtypes-obese, diabetic, and lean-are associated with varying degrees of fibrotic burden and different complications, yet the existing analytics methods often overlook its multisystem nature, intraphenotype variability, and disease dynamics. These limitations hinder accurate risk stratification and restrict personalized intervention planning.</p><p><strong>Objective: </strong>This study developed a novel, 2-stage, contrastive learning-based method to predict the phenotype of MAFLD among adults. This method leverages multiview contrastive learning; it models individual heterogeneities and important relationships in clinical and survey-based data to predict phenotypes among adults, thus supporting clinical decision-making and personalized care.</p><p><strong>Methods: </strong>Demographic, clinical, lifestyle, and genetic family history data of 4408 adults revealed how capturing essential relationships in patient data from different sources can transform individual-level representations into multiple, complementary views. Evaluation of the predictive efficacy of the proposed method in comparison with 8 prevalent methods relied on recall, precision, F1-score, and area under the curve values. Moreover, a Shapley additive explanation analysis was performed for interpretability.</p><p><strong>Results: </strong>The proposed method consistently and significantly outperformed all benchmark methods. It attained the highest F1-score, showing a 32.8% improvement for nondiabetic MAFLD (0.531 vs 0.400) and 30.4% improvement for diabetic MAFLD (0.519 vs 0.398) over the respective best-performing benchmark. The results underscore the clinical value and utility of integrating clinical and survey-based data in the prediction of MAFLD phenotypes among adults.</p><p><strong>Conclusions: </strong>The proposed method is a viable approach for MAFLD phenotype prediction. It is more effective in identifying at-risk adults than many prevalent data-driven analytics methods and thereby can enhance clinical decision-making and support patient-centric care and management.</p>","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e75747"},"PeriodicalIF":3.8,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12702840/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145758554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Riley Bove, Luca Capezzuto, Imogen West, Simon Dryden, Saira Ghafur, Jack Halligan, Stanislas Hubeaux, Agne Kazlauskaite
<p><strong>Background: </strong>Efforts are being made to integrate digital health technologies into clinical care for multiple sclerosis (MS) to improve patient monitoring. Efficiently probing how they might impact clinical care could streamline digital tool development. The Floodlight digital tool, comprising 5 smartphone sensor-based tests, was used to generate health-related data on patient function and symptoms in a clinical simulation.</p><p><strong>Objective: </strong>The study had 3 objectives: (1) assess the utility of simulated clinical encounters as a research methodology for exploring the introduction of digital health technologies into clinical practice in MS, (2) confirm the fidelity of the simulated environment and patient cases developed and understand what metrics (eg, workflow, comprehensive evaluation) could be generated, and (3) generate insights into the utility of digitally collected data, including usability, clinical decision contribution, and impact on workflows, in clinical practice.</p><p><strong>Methods: </strong>A total of 2 patient cases consisting of clinical, radiological, and digital health data were developed with clinician input. US-based neurologists prepared for and conducted 2 simulated teleconsultations each, with an actor briefed on case profiles. Floodlight data were available, via the Floodlight MS™ Health Care Professional Portal, for 1 of the 2 consultations. Participant neurologists completed interviews and surveys assessing the fidelity of the cases presented, user experience and workflow metrics, patient concerns identified, care decisions made, and confidence in making decisions.</p><p><strong>Results: </strong>All 10 neurologists indicated that the simulations were high-fidelity representations of real consultations. Using the Floodlight technology for the first time, median time taken to prepare for and conduct the consultation was ~1.7-2 minutes longer, with slightly greater mental effort reported by participants, compared with not using the tool. The Floodlight MS Health Care Professional Portal scored an "above average" 79 on the System Usability Scale and an "acceptable" Net Promoter Score of 10. In total, 6 of the 10 neurologists "strongly agreed" that it was easier and quicker to identify patient concerns when they had access to the patient-generated Floodlight data to prepare for their encounters than when they did not. Overall, more care and management decisions were taken when the digital tool was used (37 vs 29). Of those 37 decisions, Floodlight data were reported as a trigger for 20 decisions, always in combination with other elements including patient history (20/20) and clinical exam findings (9/20).</p><p><strong>Conclusions: </strong>These findings advance our understanding of clinical simulation as a method for evaluating digital tools and other innovative technologies for MS care. High-fidelity patient cases could be provided for the mock teleconsultations, and the simulated clin
背景:人们正在努力将数字健康技术整合到多发性硬化症(MS)的临床护理中,以改善患者的监测。有效地探索它们如何影响临床护理可以简化数字工具的开发。泛光灯数字工具包括5个基于智能手机传感器的测试,用于在临床模拟中生成有关患者功能和症状的健康相关数据。目的:本研究有3个目的:(1)评估模拟临床接触的效用,作为探索将数字健康技术引入MS临床实践的研究方法;(2)确认模拟环境和患者病例的保真度,并了解可以产生哪些指标(例如,工作流,综合评估);(3)对数字收集数据的效用产生见解,包括可用性、临床决策贡献和对工作流程的影响。临床实践中。方法:在临床医生的输入下,编制2例患者的临床、放射学和数字健康资料。美国的神经学家准备并进行了两次模拟远程会诊,并向一名演员简要介绍了病例概况。2次咨询中的1次可通过Floodlight MS™医疗保健专业门户获得泛光灯数据。参与的神经科医生完成了访谈和调查,评估了所呈现病例的保真度、用户体验和工作流程指标、确定的患者关注点、做出的护理决策以及做出决策的信心。结果:所有10名神经科医生表示,模拟是真实咨询的高保真表现。第一次使用泛光灯技术,准备和进行咨询的平均时间约为1.7-2分钟,与不使用该工具相比,参与者报告的心理努力略大。泛光灯MS医疗保健专业门户网站在系统可用性量表上获得了79分“高于平均水平”,净推荐值为10分,“可接受”。总的来说,10位神经科医生中有6位“强烈同意”,当他们可以访问患者生成的泛光灯数据以为他们的遭遇做准备时,比没有访问时更容易、更快地识别患者的担忧。总体而言,当使用数字工具时,采取了更多的谨慎和管理决策(37 vs 29)。在这37项决策中,泛光灯数据被报告为20项决策的触发因素,这些决策总是与其他因素相结合,包括患者病史(20/20)和临床检查结果(9/20)。结论:这些发现促进了我们对临床模拟作为评估MS护理的数字工具和其他创新技术的方法的理解。可以为模拟远程会诊提供高保真的患者病例,模拟临床环境有助于评估新的数字工具的可用性和实用性,从而为神经科医生如何访问和利用数字数据来支持常规MS护理提供初步证据。
{"title":"Simulation of Clinical Visits as a Novel Approach to Evaluate Digital Health in Multiple Sclerosis: Simulation Study.","authors":"Riley Bove, Luca Capezzuto, Imogen West, Simon Dryden, Saira Ghafur, Jack Halligan, Stanislas Hubeaux, Agne Kazlauskaite","doi":"10.2196/67845","DOIUrl":"10.2196/67845","url":null,"abstract":"<p><strong>Background: </strong>Efforts are being made to integrate digital health technologies into clinical care for multiple sclerosis (MS) to improve patient monitoring. Efficiently probing how they might impact clinical care could streamline digital tool development. The Floodlight digital tool, comprising 5 smartphone sensor-based tests, was used to generate health-related data on patient function and symptoms in a clinical simulation.</p><p><strong>Objective: </strong>The study had 3 objectives: (1) assess the utility of simulated clinical encounters as a research methodology for exploring the introduction of digital health technologies into clinical practice in MS, (2) confirm the fidelity of the simulated environment and patient cases developed and understand what metrics (eg, workflow, comprehensive evaluation) could be generated, and (3) generate insights into the utility of digitally collected data, including usability, clinical decision contribution, and impact on workflows, in clinical practice.</p><p><strong>Methods: </strong>A total of 2 patient cases consisting of clinical, radiological, and digital health data were developed with clinician input. US-based neurologists prepared for and conducted 2 simulated teleconsultations each, with an actor briefed on case profiles. Floodlight data were available, via the Floodlight MS™ Health Care Professional Portal, for 1 of the 2 consultations. Participant neurologists completed interviews and surveys assessing the fidelity of the cases presented, user experience and workflow metrics, patient concerns identified, care decisions made, and confidence in making decisions.</p><p><strong>Results: </strong>All 10 neurologists indicated that the simulations were high-fidelity representations of real consultations. Using the Floodlight technology for the first time, median time taken to prepare for and conduct the consultation was ~1.7-2 minutes longer, with slightly greater mental effort reported by participants, compared with not using the tool. The Floodlight MS Health Care Professional Portal scored an \"above average\" 79 on the System Usability Scale and an \"acceptable\" Net Promoter Score of 10. In total, 6 of the 10 neurologists \"strongly agreed\" that it was easier and quicker to identify patient concerns when they had access to the patient-generated Floodlight data to prepare for their encounters than when they did not. Overall, more care and management decisions were taken when the digital tool was used (37 vs 29). Of those 37 decisions, Floodlight data were reported as a trigger for 20 decisions, always in combination with other elements including patient history (20/20) and clinical exam findings (9/20).</p><p><strong>Conclusions: </strong>These findings advance our understanding of clinical simulation as a method for evaluating digital tools and other innovative technologies for MS care. High-fidelity patient cases could be provided for the mock teleconsultations, and the simulated clin","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e67845"},"PeriodicalIF":3.8,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12697915/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145745868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ya-Han Hu, Yi-Ying Cheng, Chung-Ching Lan, Yu-Hsiang Su, Sheng-Feng Sung
<p><strong>Background: </strong>Clinical trial eligibility screening using electronic medical records (EMRs) is challenging due to the complexity of patient data and the varied clinical terminologies. Manual screening is time-consuming, requires specialized knowledge, and can lead to inconsistent participant selection, potentially compromising patient safety and research outcomes. This is critical in time-sensitive conditions like acute ischemic stroke. While computerized clinical decision support tools offer solutions, most require software engineering expertise to update, limiting their practical utility when eligibility criteria change.</p><p><strong>Objective: </strong>We developed and evaluated the intelligent trial eligibility screening tool (iTEST), which combines natural language processing with a block-based visual programming interface designed to enable clinicians to create and modify eligibility screening rules independently. In this study, we assessed iTEST's rule evaluation module using pre-configured rules and compared its effectiveness with that of standard EMR interfaces.</p><p><strong>Methods: </strong>We conducted an experiment at a tertiary teaching hospital in Taiwan with 12 clinicians using a 2-period crossover design. The clinicians assessed the eligibility of 4 patients with stroke for 2 clinical trials using both standard EMR and iTEST in a counterbalanced order, resulting in 48 evaluation scenarios. The iTEST comprised a rule authoring module using Google Blockly and a rule evaluation module utilizing MetaMap Lite for extracting medical concepts from unstructured EMR documents and structured laboratory data. Primary outcomes included accuracy in determining eligibility. Secondary outcomes measured task completion time, cognitive workload using the National Aeronautics and Space Administration Task Load Index scale (range 0-100, with lower scores indicating a lower cognitive workload), and system usability through the system usability scale (range: 0-100, with higher scores indicating higher system usability).</p><p><strong>Results: </strong>The iTEST significantly improved accuracy scores (from 0.91 to 1.00, P<.001) and reduced completion time (from 3.18 to 2.44 min, P=.004) compared to the standard EMR interface. Users reported lower cognitive workload (National Aeronautics and Space Administration Task Load Index scale, 39.7 vs 62.8, P=.02) and higher system usability scale scores (71.3 vs 46.3, P=.01) with the iTEST. Particularly notable improvements in perceived cognitive workload were observed in temporal demand, effort, and frustration levels.</p><p><strong>Conclusions: </strong>The iTEST demonstrated superior performance in clinical trial eligibility screening, delivering improved accuracy, reduced completion time, lower cognitive workload, and better usability when evaluating preconfigured eligibility rules. The improved accuracy is critical for patient safety, as the misidentification of eligibility criteria cou
{"title":"An Intelligent Trial Eligibility Screening Tool Using Natural Language Processing With a Block-Based Visual Programming Interface: Development and Usability Study.","authors":"Ya-Han Hu, Yi-Ying Cheng, Chung-Ching Lan, Yu-Hsiang Su, Sheng-Feng Sung","doi":"10.2196/80072","DOIUrl":"10.2196/80072","url":null,"abstract":"<p><strong>Background: </strong>Clinical trial eligibility screening using electronic medical records (EMRs) is challenging due to the complexity of patient data and the varied clinical terminologies. Manual screening is time-consuming, requires specialized knowledge, and can lead to inconsistent participant selection, potentially compromising patient safety and research outcomes. This is critical in time-sensitive conditions like acute ischemic stroke. While computerized clinical decision support tools offer solutions, most require software engineering expertise to update, limiting their practical utility when eligibility criteria change.</p><p><strong>Objective: </strong>We developed and evaluated the intelligent trial eligibility screening tool (iTEST), which combines natural language processing with a block-based visual programming interface designed to enable clinicians to create and modify eligibility screening rules independently. In this study, we assessed iTEST's rule evaluation module using pre-configured rules and compared its effectiveness with that of standard EMR interfaces.</p><p><strong>Methods: </strong>We conducted an experiment at a tertiary teaching hospital in Taiwan with 12 clinicians using a 2-period crossover design. The clinicians assessed the eligibility of 4 patients with stroke for 2 clinical trials using both standard EMR and iTEST in a counterbalanced order, resulting in 48 evaluation scenarios. The iTEST comprised a rule authoring module using Google Blockly and a rule evaluation module utilizing MetaMap Lite for extracting medical concepts from unstructured EMR documents and structured laboratory data. Primary outcomes included accuracy in determining eligibility. Secondary outcomes measured task completion time, cognitive workload using the National Aeronautics and Space Administration Task Load Index scale (range 0-100, with lower scores indicating a lower cognitive workload), and system usability through the system usability scale (range: 0-100, with higher scores indicating higher system usability).</p><p><strong>Results: </strong>The iTEST significantly improved accuracy scores (from 0.91 to 1.00, P<.001) and reduced completion time (from 3.18 to 2.44 min, P=.004) compared to the standard EMR interface. Users reported lower cognitive workload (National Aeronautics and Space Administration Task Load Index scale, 39.7 vs 62.8, P=.02) and higher system usability scale scores (71.3 vs 46.3, P=.01) with the iTEST. Particularly notable improvements in perceived cognitive workload were observed in temporal demand, effort, and frustration levels.</p><p><strong>Conclusions: </strong>The iTEST demonstrated superior performance in clinical trial eligibility screening, delivering improved accuracy, reduced completion time, lower cognitive workload, and better usability when evaluating preconfigured eligibility rules. The improved accuracy is critical for patient safety, as the misidentification of eligibility criteria cou","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e80072"},"PeriodicalIF":3.8,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12698033/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145745925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sara Alessandra Keller, Maximilian Schuessler, Behzad Naderalvojoud, Tina Seto, Lu Tian, Mohana Roy, Tina Hernandez-Boussard
Background: Acute care use (ACU) represents a major economic burden in oncology, which can ideally be prevented. Existing models effectively predict such events.
Objective: We aimed to quantify the cost savings achieved by implementing a model to predict ACU in oncology patients undergoing systemic therapy.
Methods: This retrospective cohort study analyzed patients with cancer at an academic medical center from 2010 to 2022. We included patients who received systemic therapy and identified ACU events occurring after treatment initiation, excluding those with known death dates within the study period. Data on ACU-related expenses were gathered from Medicare claims and mapped to service codes in electronic health records, yielding average daily costs for each patient over 180 days following the start of therapy. The exposure was an ACU event.
Results: The main outcome was the average daily cost per patient at the end of the first 180 days of systemic therapy. We observed that expense accumulation flattened earlier and more rapidly among non-ACU patients. This study included 20,556 patients, of whom 3820 (18.58%) experienced at least 1 ACU. The average daily cost per patient for those with and without ACU was US $94.62 (SD US $72.54; 95% CI US $92.32-$96.92) and US $53.28 (SD US $59.92; 95% CI US $52.37-$54.19), respectively. The average total cost per ACU and non-ACU patient was US $17,031.92 (SD US $13,056.63; 95% CI US $16,616.74-$17,445.09) and US $9591.06 (SD US $10,785.83; 95% CI US $9427.64-$9754.48), respectively. To estimate the long-term financial impact of deploying the predictive model, we conducted a cost-benefit analysis based on an annual cohort size of 2177 patients. In the first year alone, the model yielded projected savings of US $910,000. By year 6, projected savings grew to US $9.46 million annually. The cumulative avoided costs over a 6-year deployment period totaled approximately US $31.11 million. These estimates compared the baseline cost model to the intervention model assuming a prevention rate of 35% for preventable ACU events and an average ACU cost of US $17,031.92 (SD US $13,037).
Conclusions: Predictive analytics can significantly reduce costs associated with ACU events, enhancing economic efficiency in cancer care. Further research is needed to explore potential health benefits.
背景:急性护理使用(ACU)是肿瘤学的主要经济负担,理想情况下是可以预防的。现有的模型可以有效地预测这类事件。目的:我们旨在量化通过实施预测肿瘤患者接受全身治疗的ACU模型所实现的成本节约。方法:本回顾性队列研究分析了2010年至2022年在某学术医疗中心就诊的癌症患者。我们纳入了接受全身治疗并确定在治疗开始后发生ACU事件的患者,排除了研究期间已知死亡日期的患者。从医疗保险索赔中收集了与acu相关的费用数据,并将其映射到电子健康记录中的服务代码中,得出每位患者在治疗开始后180天内的平均每日费用。暴露是ACU造成的。结果:主要观察指标为每位患者在前180天全身治疗结束时的平均每日费用。我们观察到,在非acu患者中,费用积累更早、更快地趋于平缓。本研究纳入20,556例患者,其中3820例(18.58%)经历过至少1次ACU。有无ACU患者的平均每日费用分别为94.62美元(SD $72.54; 95% CI为92.32- 96.92美元)和53.28美元(SD $59.92; 95% CI为52.37- 54.19美元)。ACU和非ACU患者的平均总成本分别为17031.92美元(标准差13,056.63美元;95% CI 16,616.74美元- 17,445.09美元)和9591.06美元(标准差10,785.83美元;95% CI 9427.64美元- 9754.48美元)。为了评估部署预测模型的长期财务影响,我们基于每年2177例患者的队列规模进行了成本效益分析。仅在第一年,该模式就预计节省了91万美元。到第6年,预计每年节余增加到946万美元。在6年的部署期间,累计避免的费用总计约为3111万美元。这些估计将基线成本模型与干预模型进行了比较,假设可预防的ACU事件的预防率为35%,平均ACU成本为17031.92美元(SD $13,037)。结论:预测分析可以显著降低与ACU事件相关的成本,提高癌症治疗的经济效率。需要进一步的研究来探索其潜在的健康益处。
{"title":"Cost-Benefit Analysis of Preventing Acute Care Use in Oncology Patients Following Systemic Therapy Using Medicare Claims Data: Retrospective Cohort Study.","authors":"Sara Alessandra Keller, Maximilian Schuessler, Behzad Naderalvojoud, Tina Seto, Lu Tian, Mohana Roy, Tina Hernandez-Boussard","doi":"10.2196/77891","DOIUrl":"10.2196/77891","url":null,"abstract":"<p><strong>Background: </strong>Acute care use (ACU) represents a major economic burden in oncology, which can ideally be prevented. Existing models effectively predict such events.</p><p><strong>Objective: </strong>We aimed to quantify the cost savings achieved by implementing a model to predict ACU in oncology patients undergoing systemic therapy.</p><p><strong>Methods: </strong>This retrospective cohort study analyzed patients with cancer at an academic medical center from 2010 to 2022. We included patients who received systemic therapy and identified ACU events occurring after treatment initiation, excluding those with known death dates within the study period. Data on ACU-related expenses were gathered from Medicare claims and mapped to service codes in electronic health records, yielding average daily costs for each patient over 180 days following the start of therapy. The exposure was an ACU event.</p><p><strong>Results: </strong>The main outcome was the average daily cost per patient at the end of the first 180 days of systemic therapy. We observed that expense accumulation flattened earlier and more rapidly among non-ACU patients. This study included 20,556 patients, of whom 3820 (18.58%) experienced at least 1 ACU. The average daily cost per patient for those with and without ACU was US $94.62 (SD US $72.54; 95% CI US $92.32-$96.92) and US $53.28 (SD US $59.92; 95% CI US $52.37-$54.19), respectively. The average total cost per ACU and non-ACU patient was US $17,031.92 (SD US $13,056.63; 95% CI US $16,616.74-$17,445.09) and US $9591.06 (SD US $10,785.83; 95% CI US $9427.64-$9754.48), respectively. To estimate the long-term financial impact of deploying the predictive model, we conducted a cost-benefit analysis based on an annual cohort size of 2177 patients. In the first year alone, the model yielded projected savings of US $910,000. By year 6, projected savings grew to US $9.46 million annually. The cumulative avoided costs over a 6-year deployment period totaled approximately US $31.11 million. These estimates compared the baseline cost model to the intervention model assuming a prevention rate of 35% for preventable ACU events and an average ACU cost of US $17,031.92 (SD US $13,037).</p><p><strong>Conclusions: </strong>Predictive analytics can significantly reduce costs associated with ACU events, enhancing economic efficiency in cancer care. Further research is needed to explore potential health benefits.</p>","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e77891"},"PeriodicalIF":3.8,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12698071/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145745862","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Adhari Al Zaabi, Rashid Alshibli, Abdullah AlAmri, Ibrahim AlRuheili, Syaheerah Lebai Lutfi
Background: The use of large language models (LLMs) in radiology is expanding rapidly, offering new possibilities in report generation, decision support, and workflow optimization. However, a comprehensive evaluation of their applications, performance, and limitations across the radiology domain remains limited.
Objective: This review aimed to map current applications of LLMs in radiology, evaluate their performance across key tasks, and identify prevailing limitations and directions for future research.
Methods: A scoping review was conducted in accordance with the framework by Arksey and O'Malley framework and the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines. Three databases-PubMed, ScopusCOPUS, and IEEE Xplore-were searched for peer-reviewed studies published between January 2022 and December 2024. Eligible studies included empirical evaluations of LLMs applied to radiological data or workflows. Commentaries, reviews, and technical model proposals without evaluation were excluded. Two reviewers independently screened studies and extracted data on study characteristics, LLM type, radiological use case, data modality, and evaluation metrics. A thematic synthesis was used to identify key domains of application. No formal risk-of-bias assessment was performed, but a narrative appraisal of dataset representativeness and study quality was included.
Results: A total of 67 studies were included. (n/N, %)GPT-4 was the most frequently used model (n=28, 42%), with text-based corpora as the primary type of data used (n=43, 64%). Identified use cases fell into three thematic domains: (1) decision support (n=39, 58%), (2) report generation and summarization (n=16, 24%), and (3) workflow optimization (n=12, 18%). While LLMs demonstrated strong performance in structured-text tasks (eg, report simplification with >94% accuracy), diagnostic performance varied widely (16%-86%) and was limited by dataset bias, lack of fine tuning, and minimal clinical validation. Most studies (n=53, 79.1%) had single-center, proof-of-concept designs with limited generalizability.
Conclusions: LLMs show strong potential for augmenting radiological workflows, particularly for structured reporting, summarization, and educational tasks. However, their diagnostic performance remains inconsistent, and current implementations lack robust external validation. Future work should prioritize prospective, multicenter validation of domain-adapted and multimodal models to support safe clinical integration.
{"title":"Trends and Trajectories in the Rise of Large Language Models in Radiology: Scoping Review.","authors":"Adhari Al Zaabi, Rashid Alshibli, Abdullah AlAmri, Ibrahim AlRuheili, Syaheerah Lebai Lutfi","doi":"10.2196/78041","DOIUrl":"10.2196/78041","url":null,"abstract":"<p><strong>Background: </strong>The use of large language models (LLMs) in radiology is expanding rapidly, offering new possibilities in report generation, decision support, and workflow optimization. However, a comprehensive evaluation of their applications, performance, and limitations across the radiology domain remains limited.</p><p><strong>Objective: </strong>This review aimed to map current applications of LLMs in radiology, evaluate their performance across key tasks, and identify prevailing limitations and directions for future research.</p><p><strong>Methods: </strong>A scoping review was conducted in accordance with the framework by Arksey and O'Malley framework and the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines. Three databases-PubMed, ScopusCOPUS, and IEEE Xplore-were searched for peer-reviewed studies published between January 2022 and December 2024. Eligible studies included empirical evaluations of LLMs applied to radiological data or workflows. Commentaries, reviews, and technical model proposals without evaluation were excluded. Two reviewers independently screened studies and extracted data on study characteristics, LLM type, radiological use case, data modality, and evaluation metrics. A thematic synthesis was used to identify key domains of application. No formal risk-of-bias assessment was performed, but a narrative appraisal of dataset representativeness and study quality was included.</p><p><strong>Results: </strong>A total of 67 studies were included. (n/N, %)GPT-4 was the most frequently used model (n=28, 42%), with text-based corpora as the primary type of data used (n=43, 64%). Identified use cases fell into three thematic domains: (1) decision support (n=39, 58%), (2) report generation and summarization (n=16, 24%), and (3) workflow optimization (n=12, 18%). While LLMs demonstrated strong performance in structured-text tasks (eg, report simplification with >94% accuracy), diagnostic performance varied widely (16%-86%) and was limited by dataset bias, lack of fine tuning, and minimal clinical validation. Most studies (n=53, 79.1%) had single-center, proof-of-concept designs with limited generalizability.</p><p><strong>Conclusions: </strong>LLMs show strong potential for augmenting radiological workflows, particularly for structured reporting, summarization, and educational tasks. However, their diagnostic performance remains inconsistent, and current implementations lack robust external validation. Future work should prioritize prospective, multicenter validation of domain-adapted and multimodal models to support safe clinical integration.</p>","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e78041"},"PeriodicalIF":3.8,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12688054/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145716933","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background: Liver failure often results in significant coagulation dysfunction, which is a major complication. Artificial liver support systems (ALSS) have been used to ameliorate coagulation parameters, but the dynamic nature of these improvements and the development of predictive models remain insufficiently explored.
Objective: This study aimed to evaluate the effects of ALSS on coagulation function and to develop a dynamic prediction model using machine learning techniques to predict the improvement trends of coagulation parameters.
Methods: A systematic search was conducted in PubMed, Embase, and other databases to identify relevant studies, resulting in 18 studies comprising 1771 patients. A meta-analysis was performed to assess the impact of ALSS on coagulation parameters, including international normalized ratio (INR), prothrombin time (PT), activated partial thromboplastin time (APTT), and fibrinogen levels. In addition, clinical data from the Medical Information Mart for Intensive Care database were used to construct prediction models using logistic regression, extreme gradient boosting, random forest, and long short-term memory networks.
Results: Meta-analysis results showed that ALSS significantly improved INR, PT, APTT, and fibrinogen levels (all P<.05), with the treatment efficacy varying by modality. Among the machine learning models, the random forest model demonstrated the best performance, achieving an area under the curve of 92.12%. Dynamic INR was identified as the key predictor for coagulation abnormalities.
Conclusions: This study systematically evaluated the effects of ALSS on coagulation function in patients with liver failure, demonstrating significant improvements in key parameters such as INR, PT, and APTT, with efficacy varying across different treatment modalities. Simultaneously, a machine learning model built using intensive care unit clinical data exhibited strong predictive capability for identifying the risk of coagulation dysfunction, particularly useful in supporting early-stage clinical recognition of high-risk patients and guiding personalized coagulation management strategies. It is important to emphasize that this model is positioned as a dynamic risk alert and assessment tool, intended to assist clinical baseline evaluation and nursing interventions, rather than serving as direct validation of ALSS therapeutic efficacy.
{"title":"Coagulation Risk Prediction in Patients With Liver Failure: Integrated Meta-Analysis and Machine Learning Model Study.","authors":"Hao Wang, Tao He, Liang Ren, Tingjun Zhang","doi":"10.2196/76348","DOIUrl":"10.2196/76348","url":null,"abstract":"<p><strong>Background: </strong>Liver failure often results in significant coagulation dysfunction, which is a major complication. Artificial liver support systems (ALSS) have been used to ameliorate coagulation parameters, but the dynamic nature of these improvements and the development of predictive models remain insufficiently explored.</p><p><strong>Objective: </strong>This study aimed to evaluate the effects of ALSS on coagulation function and to develop a dynamic prediction model using machine learning techniques to predict the improvement trends of coagulation parameters.</p><p><strong>Methods: </strong>A systematic search was conducted in PubMed, Embase, and other databases to identify relevant studies, resulting in 18 studies comprising 1771 patients. A meta-analysis was performed to assess the impact of ALSS on coagulation parameters, including international normalized ratio (INR), prothrombin time (PT), activated partial thromboplastin time (APTT), and fibrinogen levels. In addition, clinical data from the Medical Information Mart for Intensive Care database were used to construct prediction models using logistic regression, extreme gradient boosting, random forest, and long short-term memory networks.</p><p><strong>Results: </strong>Meta-analysis results showed that ALSS significantly improved INR, PT, APTT, and fibrinogen levels (all P<.05), with the treatment efficacy varying by modality. Among the machine learning models, the random forest model demonstrated the best performance, achieving an area under the curve of 92.12%. Dynamic INR was identified as the key predictor for coagulation abnormalities.</p><p><strong>Conclusions: </strong>This study systematically evaluated the effects of ALSS on coagulation function in patients with liver failure, demonstrating significant improvements in key parameters such as INR, PT, and APTT, with efficacy varying across different treatment modalities. Simultaneously, a machine learning model built using intensive care unit clinical data exhibited strong predictive capability for identifying the risk of coagulation dysfunction, particularly useful in supporting early-stage clinical recognition of high-risk patients and guiding personalized coagulation management strategies. It is important to emphasize that this model is positioned as a dynamic risk alert and assessment tool, intended to assist clinical baseline evaluation and nursing interventions, rather than serving as direct validation of ALSS therapeutic efficacy.</p>","PeriodicalId":56334,"journal":{"name":"JMIR Medical Informatics","volume":"13 ","pages":"e76348"},"PeriodicalIF":3.8,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12723362/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145709516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}