Pub Date : 2026-02-25eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooag020
Gregory L Watson, Grace Staples, Robin Carver, Akhil Bhargava, Carlos López-Espina, Lee Schmalz, Farhan Ali, Peter S Antkowiak, Saleem Azad, Ramona Berghea, Lavneet Chawla, Matthew Crisp, Alon Dagan, Francisco Davila, Hugo Davila, Carmen DeMarco, Amanda Doodlesack, Aimee Espinosa, Neil S Evans, Clinton Ezekiel, Andrew Friederich, Falgun Gosai, Alexandra Halalau, Karthik Iyer, Max S Kravitz, Niko Kurtzman, John H Lee, Nicholas Maddens, Roneil Malkani, Stockton Mayer, Vikram Oke, Ashok V Palagiri, Roshni Patel, Lekshminarayan Raghavakurup, Samuel Raouf, Eric Reseland, Farid Sadaka, Deesha Sarma, Scott Smith, Tatyana Shvilkina, Matthew D Sims, Sahib Singh, Bryan A Stenson, Anwaruddin Syed, Muleta Tafa, Kurian Thomas, Sihai Dave Zhao, Ruoqing Zhu, Rashid Bashir, Bobby Reddy, Nathan I Shapiro
Objectives: To assess the interpretability and acceptance of Shapley values for making artificial intelligence/machine learning (AI/ML) tools more transparent, interpretable, and useful to clinicians.
Materials and methods: Structured assessments were conducted with 30 clinicians (15 providers; 15 nurses; 8 assessments per clinician) to evaluate their ability to understand interventional Shapley Additive exPlanations (SHAP) values, a type of Shapley value that provides individualized variable importance scores and ascertain their perspective on SHAP value utility for the use of an AI/ML sepsis diagnostic. Participants were shown the diagnostic interface for real clinical scenarios with de-identified patient data with and without SHAP values. The primary outcomes were clinician ability to correctly interpret SHAP values and clinician self-reported improvement in their understanding of how the AI/ML algorithm produced its result.
Results: Participants correctly interpreted SHAP values in 235 of 240 assessments (98%; CI, 95%-99%) and reported SHAP values improved their understanding of how the algorithm produced its result in every case (240/240; 100%; CI, 99%-100%). Participants were unanimous (30/30) in preferring the interface with SHAP values over the interface without.
Discussion: Clinician participants strongly preferred the device interface with SHAP values, were unanimous in reporting SHAP values improved their understanding of the AI/ML diagnostic, and scored nearly perfectly when asked to interpret SHAP values.
Conclusion: These results suggest health care providers value transparency into AI/ML algorithms designed for clinical use, and that Shapley values are a useful approach to providing that transparency, which in turn may improve tool adoption and clinical utility.
{"title":"Interpretability of an FDA-authorized AI/ML sepsis diagnostic tool improved by SHAP values.","authors":"Gregory L Watson, Grace Staples, Robin Carver, Akhil Bhargava, Carlos López-Espina, Lee Schmalz, Farhan Ali, Peter S Antkowiak, Saleem Azad, Ramona Berghea, Lavneet Chawla, Matthew Crisp, Alon Dagan, Francisco Davila, Hugo Davila, Carmen DeMarco, Amanda Doodlesack, Aimee Espinosa, Neil S Evans, Clinton Ezekiel, Andrew Friederich, Falgun Gosai, Alexandra Halalau, Karthik Iyer, Max S Kravitz, Niko Kurtzman, John H Lee, Nicholas Maddens, Roneil Malkani, Stockton Mayer, Vikram Oke, Ashok V Palagiri, Roshni Patel, Lekshminarayan Raghavakurup, Samuel Raouf, Eric Reseland, Farid Sadaka, Deesha Sarma, Scott Smith, Tatyana Shvilkina, Matthew D Sims, Sahib Singh, Bryan A Stenson, Anwaruddin Syed, Muleta Tafa, Kurian Thomas, Sihai Dave Zhao, Ruoqing Zhu, Rashid Bashir, Bobby Reddy, Nathan I Shapiro","doi":"10.1093/jamiaopen/ooag020","DOIUrl":"https://doi.org/10.1093/jamiaopen/ooag020","url":null,"abstract":"<p><strong>Objectives: </strong>To assess the interpretability and acceptance of Shapley values for making artificial intelligence/machine learning (AI/ML) tools more transparent, interpretable, and useful to clinicians.</p><p><strong>Materials and methods: </strong>Structured assessments were conducted with 30 clinicians (15 providers; 15 nurses; 8 assessments per clinician) to evaluate their ability to understand interventional Shapley Additive exPlanations (SHAP) values, a type of Shapley value that provides individualized variable importance scores and ascertain their perspective on SHAP value utility for the use of an AI/ML sepsis diagnostic. Participants were shown the diagnostic interface for real clinical scenarios with de-identified patient data with and without SHAP values. The primary outcomes were clinician ability to correctly interpret SHAP values and clinician self-reported improvement in their understanding of how the AI/ML algorithm produced its result.</p><p><strong>Results: </strong>Participants correctly interpreted SHAP values in 235 of 240 assessments (98%; CI, 95%-99%) and reported SHAP values improved their understanding of how the algorithm produced its result in every case (240/240; 100%; CI, 99%-100%). Participants were unanimous (30/30) in preferring the interface with SHAP values over the interface without.</p><p><strong>Discussion: </strong>Clinician participants strongly preferred the device interface with SHAP values, were unanimous in reporting SHAP values improved their understanding of the AI/ML diagnostic, and scored nearly perfectly when asked to interpret SHAP values.</p><p><strong>Conclusion: </strong>These results suggest health care providers value transparency into AI/ML algorithms designed for clinical use, and that Shapley values are a useful approach to providing that transparency, which in turn may improve tool adoption and clinical utility.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooag020"},"PeriodicalIF":3.4,"publicationDate":"2026-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12936051/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147327413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-25eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooag018
Huan Li, Varada V Khanna, Nate Apathy, A Jay Holmgren, Andrew J Loza, Edward R Melnick
Objective: To explore the relationship between ambulatory physician electronic health record (EHR) use characteristics and proxies for physician efficiency.
Materials and methods: A longitudinal cohort study was conducted to examine physician-month EHR use metadata in 413 US organizations between May 2019 and April 2022. A multi-model machine learning classifier was developed to predict physician efficiency. The main outcomes of the study were physician efficiency, measured as the proportion of same-day chart completion by specialty, and productivity, measured as daily patient visit volume, both segmented into quintiles.
Results: The study included 218 610 unique physicians with 5 193 385 physician-month observations from 413 organizations with an average chart completion efficiency of 72.9% and 10.8 visits per scheduled day. The primary ML analysis achieved an accuracy of 0.74 in classifying physician-months with high chart completion efficiency and highlighted associations with key features, such as inbox message turnaround time <1.5 days and after-hours documentation <25 min/scheduled day. A secondary analysis achieved an accuracy of 0.84 in classifying physician-months with high visit volumes, indicating that factors such as EHR time outside scheduled hours <4.1 min/visit and clinical review time <3.2 min/visit were associated with higher visit volumes.
Discussion and conclusion: Implementing specific EHR use measures with distinct thresholds, such as inbox management and after-hours documentation, could help target interventions to enhance productivity, providing actionable insights to create balanced and efficient work environments that improve patient care and reduce EHR time.
{"title":"Electronic health record use factors linked to efficiency and productivity: an explainable machine learning analysis.","authors":"Huan Li, Varada V Khanna, Nate Apathy, A Jay Holmgren, Andrew J Loza, Edward R Melnick","doi":"10.1093/jamiaopen/ooag018","DOIUrl":"https://doi.org/10.1093/jamiaopen/ooag018","url":null,"abstract":"<p><strong>Objective: </strong>To explore the relationship between ambulatory physician electronic health record (EHR) use characteristics and proxies for physician efficiency.</p><p><strong>Materials and methods: </strong>A longitudinal cohort study was conducted to examine physician-month EHR use metadata in 413 US organizations between May 2019 and April 2022. A multi-model machine learning classifier was developed to predict physician efficiency. The main outcomes of the study were physician <i>efficiency</i>, measured as the proportion of same-day chart completion by specialty, and <i>productivity</i>, measured as daily patient visit volume, both segmented into quintiles.</p><p><strong>Results: </strong>The study included 218 610 unique physicians with 5 193 385 physician-month observations from 413 organizations with an average chart completion efficiency of 72.9% and 10.8 visits per scheduled day. The primary ML analysis achieved an accuracy of 0.74 in classifying physician-months with high chart completion efficiency and highlighted associations with key features, such as inbox message turnaround time <1.5 days and after-hours documentation <25 min/scheduled day. A secondary analysis achieved an accuracy of 0.84 in classifying physician-months with high visit volumes, indicating that factors such as EHR time outside scheduled hours <4.1 min/visit and clinical review time <3.2 min/visit were associated with higher visit volumes.</p><p><strong>Discussion and conclusion: </strong>Implementing specific EHR use measures with distinct thresholds, such as inbox management and after-hours documentation, could help target interventions to enhance productivity, providing actionable insights to create balanced and efficient work environments that improve patient care and reduce EHR time.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooag018"},"PeriodicalIF":3.4,"publicationDate":"2026-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12936052/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147327450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-23eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooaf173
[This corrects the article DOI: 10.1093/jamiaopen/ooaf134.].
[这更正了文章DOI: 10.1093/jamiaopen/ooaf134.]。
{"title":"Correction to: Biomedical data repositories require governance for artificial intelligence/machine learning applications at every step.","authors":"","doi":"10.1093/jamiaopen/ooaf173","DOIUrl":"https://doi.org/10.1093/jamiaopen/ooaf173","url":null,"abstract":"<p><p>[This corrects the article DOI: 10.1093/jamiaopen/ooaf134.].</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooaf173"},"PeriodicalIF":3.4,"publicationDate":"2026-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12928734/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147285553","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-21eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooaf149
Muhammad Imran, Olga Zamaraeva, Carlos Gómez-Rodríguez
Objective: This study evaluates the usefulness of explicit syntactic knowledge, integrated via a neural mechanism, in improving the accuracy of named entity recognition in the domain of biomedical text processing.
Materials and methods: Syntactic structure of a text can be helpful to determine whether a certain part of the text is an entity or not. Parsing is an essential technique in natural language processing (NLP) that can be utilized to determine the syntactic structure of sentences in human languages. We propose to infuse syntactic knowledge through the attention mechanism using dependency parsing and sequence labelling parsing, as well as the multi-task learning paradigm. Experiments were conducted on five datasets: MTSamples, VAERS, NCBI-disease, BC2GM, and JNLPBA.
Results: We demonstrate improvements in the F1 score over the current state of the art on 3 out of 5 datasets (MTSamples, VAERS, and NCBI).
Discussion: We reduce the number of mismatches with gold labels in particular in the n-dash and parentheses tokens and in compound and adjective modifier dependencies.
Conclusion: Syntactic features improve NER accuracy in attention-based neural systems, and parsing as sequence labelling brings additional benefits.
{"title":"SynNER: syntax-infused named entity recognition in the biomedical domain.","authors":"Muhammad Imran, Olga Zamaraeva, Carlos Gómez-Rodríguez","doi":"10.1093/jamiaopen/ooaf149","DOIUrl":"10.1093/jamiaopen/ooaf149","url":null,"abstract":"<p><strong>Objective: </strong>This study evaluates the usefulness of explicit syntactic knowledge, integrated via a neural mechanism, in improving the accuracy of named entity recognition in the domain of biomedical text processing.</p><p><strong>Materials and methods: </strong>Syntactic structure of a text can be helpful to determine whether a certain part of the text is an entity or not. Parsing is an essential technique in natural language processing (NLP) that can be utilized to determine the syntactic structure of sentences in human languages. We propose to infuse syntactic knowledge through the attention mechanism using dependency parsing and sequence labelling parsing, as well as the multi-task learning paradigm. Experiments were conducted on five datasets: MTSamples, VAERS, NCBI-disease, BC2GM, and JNLPBA.</p><p><strong>Results: </strong>We demonstrate improvements in the F1 score over the current state of the art on 3 out of 5 datasets (MTSamples, VAERS, and NCBI).</p><p><strong>Discussion: </strong>We reduce the number of mismatches with gold labels in particular in the n-dash and parentheses tokens and in compound and adjective modifier dependencies.</p><p><strong>Conclusion: </strong>Syntactic features improve NER accuracy in attention-based neural systems, and parsing as sequence labelling brings additional benefits.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooaf149"},"PeriodicalIF":3.4,"publicationDate":"2026-02-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12932951/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147310891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-18eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooag019
Yang Yang, Kathryn I Pollak, Bibhas Chakraborty, Molei Liu, Doudou Zhou, Chuan Hong
Objectives: Electronic health record (EHR) phenotyping often relies on noisy proxy labels, which undermine the reliability of downstream risk prediction. Active learning can reduce annotation costs, but typical heuristics do not directly optimize downstream prediction. Our goal was to develop a framework that directly uses downstream prediction performance as feedback to guide phenotype correction and sample selection under constrained labeling budgets.
Materials and methods: We propose reinforcement-enhanced label-efficient active phenotyping (RELEAP), a reinforcement learning-based active learning framework. Reinforcement-enhanced label-efficient adaptively integrates multiple querying strategies and, unlike prior methods, updates its policy based on feedback from downstream models. We evaluated RELEAP on a de-identified Duke University Health System (DUHS) cohort (2014-2024) for incident lung cancer risk prediction, using logistic regression and penalized Cox survival models. Performance was benchmarked against noisy-label baselines and single-strategy active learning.
Results: Reinforcement-enhanced label-efficient improved over the proxy-only baseline and approached oracle performance under the same budget. Logistic AUC increased from 0.774 to 0.807. Survival concordance index increased from 0.715 to 0.749. Gains were stable across iterations using downstream feedback. These trends were consistent in sex-stratified subgroup analyses (female vs male).
Discussion: By linking phenotype refinement to prediction outcomes, RELEAP learns which samples most improve downstream discrimination and calibration, offering a more principled alternative to fixed active learning rules.
Conclusion: Reinforcement-enhanced label-efficient optimizes phenotype correction through downstream feedback, offering a scalable, label-efficient paradigm that reduces manual chart review and enhances the reliability of EHR-based risk prediction.
{"title":"RELEAP: reinforcement-enhanced label-efficient active phenotyping for electronic health records.","authors":"Yang Yang, Kathryn I Pollak, Bibhas Chakraborty, Molei Liu, Doudou Zhou, Chuan Hong","doi":"10.1093/jamiaopen/ooag019","DOIUrl":"https://doi.org/10.1093/jamiaopen/ooag019","url":null,"abstract":"<p><strong>Objectives: </strong>Electronic health record (EHR) phenotyping often relies on noisy proxy labels, which undermine the reliability of downstream risk prediction. Active learning can reduce annotation costs, but typical heuristics do not directly optimize downstream prediction. Our goal was to develop a framework that directly uses downstream prediction performance as feedback to guide phenotype correction and sample selection under constrained labeling budgets.</p><p><strong>Materials and methods: </strong>We propose reinforcement-enhanced label-efficient active phenotyping (RELEAP), a reinforcement learning-based active learning framework. Reinforcement-enhanced label-efficient adaptively integrates multiple querying strategies and, unlike prior methods, updates its policy based on feedback from downstream models. We evaluated RELEAP on a de-identified Duke University Health System (DUHS) cohort (2014-2024) for incident lung cancer risk prediction, using logistic regression and penalized Cox survival models. Performance was benchmarked against noisy-label baselines and single-strategy active learning.</p><p><strong>Results: </strong>Reinforcement-enhanced label-efficient improved over the proxy-only baseline and approached oracle performance under the same budget. Logistic AUC increased from 0.774 to 0.807. Survival concordance index increased from 0.715 to 0.749. Gains were stable across iterations using downstream feedback. These trends were consistent in sex-stratified subgroup analyses (female vs male).</p><p><strong>Discussion: </strong>By linking phenotype refinement to prediction outcomes, RELEAP learns which samples most improve downstream discrimination and calibration, offering a more principled alternative to fixed active learning rules.</p><p><strong>Conclusion: </strong>Reinforcement-enhanced label-efficient optimizes phenotype correction through downstream feedback, offering a scalable, label-efficient paradigm that reduces manual chart review and enhances the reliability of EHR-based risk prediction.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooag019"},"PeriodicalIF":3.4,"publicationDate":"2026-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12918302/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147272274","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-17eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooag016
Steven N Hart, Teya S Bergamaschi
Objectives: To develop and validate an agent-based Large Language Model (LLM) system for extracting structured data from breast cancer synoptic pathology reports and assess the performance gap between synthetic and real-world validation.
Materials and methods: We developed a modular artificial intelligence (AI) agent-based framework employing sequential specialized LLMs for parsing pathology reports and extracting structured data. We normalized College of American Pathologists (CAP) cancer protocols into 8 sections, 86 subsections, and 229 discrete fields. Seven leading LLMs (gemini-2.5-pro, llama3.3-70b, phi4-14b, deepseek-r1 14B/70B, gemma3-27b, gemini-2.0-flash-lite) were validated using dual evaluation: synthetic validation (864 controlled test cases) and real-world ground truth (6651 annotated fields from 90 pathology reports).
Results: Synthetic validation demonstrated strong performance (accuracy: 93.8%-99.0%). Real-world evaluation revealed field extraction recall ranging from 61.8% to 87.7%, demonstrating a substantial "reality gap" with performance drops of 11-32 percentage points. The gemini-2.5-pro model achieved the highest real-world recall (87.7%). Model size did not predict performance: the 14B-parameter deepseek-r1 (77.6%) outperformed its 70B-parameter counterpart (70.4%).
Discussion: The substantial performance degradation from synthetic to real-world data underscores the complexity of authentic clinical documentation. Smaller models can achieve competitive or superior recall, reducing computational costs. With even the best models missing 12%-38% of annotated fields, mandatory human verification is essential for clinical deployment.
Conclusion: While LLM-based extraction systems show promise for pathology data extraction, synthetic validation alone provides false confidence. Rigorous real-world ground truth evaluation with expert annotation is essential before clinical deployment. These systems are best positioned as screening tools with mandatory human oversight rather than autonomous decision-making systems.
{"title":"Agent-based large language model system for extracting structured data from breast cancer synoptic reports: a dual-validation study.","authors":"Steven N Hart, Teya S Bergamaschi","doi":"10.1093/jamiaopen/ooag016","DOIUrl":"10.1093/jamiaopen/ooag016","url":null,"abstract":"<p><strong>Objectives: </strong>To develop and validate an agent-based Large Language Model (LLM) system for extracting structured data from breast cancer synoptic pathology reports and assess the performance gap between synthetic and real-world validation.</p><p><strong>Materials and methods: </strong>We developed a modular artificial intelligence (AI) agent-based framework employing sequential specialized LLMs for parsing pathology reports and extracting structured data. We normalized College of American Pathologists (CAP) cancer protocols into 8 sections, 86 subsections, and 229 discrete fields. Seven leading LLMs (gemini-2.5-pro, llama3.3-70b, phi4-14b, deepseek-r1 14B/70B, gemma3-27b, gemini-2.0-flash-lite) were validated using dual evaluation: synthetic validation (864 controlled test cases) and real-world ground truth (6651 annotated fields from 90 pathology reports).</p><p><strong>Results: </strong>Synthetic validation demonstrated strong performance (accuracy: 93.8%-99.0%). Real-world evaluation revealed field extraction recall ranging from 61.8% to 87.7%, demonstrating a substantial \"reality gap\" with performance drops of 11-32 percentage points. The gemini-2.5-pro model achieved the highest real-world recall (87.7%). Model size did not predict performance: the 14B-parameter deepseek-r1 (77.6%) outperformed its 70B-parameter counterpart (70.4%).</p><p><strong>Discussion: </strong>The substantial performance degradation from synthetic to real-world data underscores the complexity of authentic clinical documentation. Smaller models can achieve competitive or superior recall, reducing computational costs. With even the best models missing 12%-38% of annotated fields, mandatory human verification is essential for clinical deployment.</p><p><strong>Conclusion: </strong>While LLM-based extraction systems show promise for pathology data extraction, synthetic validation alone provides false confidence. Rigorous real-world ground truth evaluation with expert annotation is essential before clinical deployment. These systems are best positioned as screening tools with mandatory human oversight rather than autonomous decision-making systems.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooag016"},"PeriodicalIF":3.4,"publicationDate":"2026-02-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12932940/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147310932","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-17eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooag021
Eilon Gabel, Stephen Murray, Tristan Grogan, Theodora Wingert, Ira Hofer
Background: Traditional clinical trial enrollment relies on manual screening and coordinator-led recruitment, creating scalability barriers in high-volume perioperative environments. This study evaluated whether a fully automated, electronic health record (EHR)-integrated clinical decision support (CDS) system could identify eligible patients and engage clinicians in real time without manual screening or dedicated research staff.
Methods: In this prospective implementation study, predefined respiratory-risk criteria were computed within the UCLA Perioperative Data Warehouse and transmitted to the EHR via Healthcare Level Seven interfaces. Patients meeting inclusion criteria automatically triggered Best Practice Advisories (BPAs) recommending an intervention. Outcomes included system accuracy in eligibility identification, provider adherence to BPA recommendations, and technical performance metrics.
Results: The automated system processed 10 592 eligible patients and achieved 51.2% provider adherence (5424 patients) to CDS prompts without coordinator involvement. BPA allocation accuracy was 69.7% among patients recovering in the post-anesthesia care unit and 59.4% when including unanticipated ICU transfers. Adherence varied significantly by care team composition, with full teams (attending + CRNA + resident) achieving 57.4% adherence compared with 42.2% for solo attendings. Workflow factors were stronger predictors of adherence than patient clinical characteristics, indicating minimal selection bias.
Conclusions: Fully automated, EHR-integrated CDS can enable large-scale, workflow-embedded enrollment into implementation-focused studies. While not a substitute for research designs requiring consent or randomization, this framework demonstrates a scalable approach for automated prescreening and CDS-driven prompting that reduces reliance on coordinator-dependent processes and supports real-world implementation science.
背景:传统的临床试验招募依赖于人工筛选和协调员主导的招募,这在大容量围手术期环境中造成了可扩展性障碍。本研究评估了一个完全自动化的电子健康记录(EHR)集成临床决策支持(CDS)系统是否可以识别符合条件的患者,并在没有人工筛查或专门研究人员的情况下实时吸引临床医生。方法:在这项前瞻性实施研究中,在加州大学洛杉矶分校围手术期数据仓库中计算预定义的呼吸风险标准,并通过医疗保健级别7接口传输到电子病历。符合纳入标准的患者会自动触发最佳实践建议(BPAs),建议进行干预。结果包括系统在资格识别方面的准确性,供应商对BPA建议的依从性,以及技术性能指标。结果:自动化系统处理了10 592名符合条件的患者,在没有协调员参与的情况下,提供者对CDS提示的依从性达到51.2%(5424名患者)。在麻醉后护理病房康复的患者中,BPA分配准确率为69.7%,在包括意外ICU转移的患者中,BPA分配准确率为59.4%。依从性因护理团队组成而有显著差异,全队(主治医师+ CRNA +住院医师)的依从性为57.4%,而单独主治医师的依从性为42.2%。工作流程因素比患者临床特征更能预测依从性,表明选择偏差最小。结论:完全自动化、ehr集成的CDS可以使大规模、嵌入工作流程的入组成为以实施为重点的研究。虽然不能替代需要同意或随机化的研究设计,但该框架展示了一种可扩展的自动预筛选和cd驱动提示方法,减少了对协调员依赖过程的依赖,并支持现实世界的实施科学。
{"title":"Automating eligibility assessment and enrollment for sugammadex administration within an integrated perioperative workflow.","authors":"Eilon Gabel, Stephen Murray, Tristan Grogan, Theodora Wingert, Ira Hofer","doi":"10.1093/jamiaopen/ooag021","DOIUrl":"10.1093/jamiaopen/ooag021","url":null,"abstract":"<p><strong>Background: </strong>Traditional clinical trial enrollment relies on manual screening and coordinator-led recruitment, creating scalability barriers in high-volume perioperative environments. This study evaluated whether a fully automated, electronic health record (EHR)-integrated clinical decision support (CDS) system could identify eligible patients and engage clinicians in real time without manual screening or dedicated research staff.</p><p><strong>Methods: </strong>In this prospective implementation study, predefined respiratory-risk criteria were computed within the UCLA Perioperative Data Warehouse and transmitted to the EHR via Healthcare Level Seven interfaces. Patients meeting inclusion criteria automatically triggered Best Practice Advisories (BPAs) recommending an intervention. Outcomes included system accuracy in eligibility identification, provider adherence to BPA recommendations, and technical performance metrics.</p><p><strong>Results: </strong>The automated system processed 10 592 eligible patients and achieved 51.2% provider adherence (5424 patients) to CDS prompts without coordinator involvement. BPA allocation accuracy was 69.7% among patients recovering in the post-anesthesia care unit and 59.4% when including unanticipated ICU transfers. Adherence varied significantly by care team composition, with full teams (attending + CRNA + resident) achieving 57.4% adherence compared with 42.2% for solo attendings. Workflow factors were stronger predictors of adherence than patient clinical characteristics, indicating minimal selection bias.</p><p><strong>Conclusions: </strong>Fully automated, EHR-integrated CDS can enable large-scale, workflow-embedded enrollment into implementation-focused studies. While not a substitute for research designs requiring consent or randomization, this framework demonstrates a scalable approach for automated prescreening and CDS-driven prompting that reduces reliance on coordinator-dependent processes and supports real-world implementation science.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooag021"},"PeriodicalIF":3.4,"publicationDate":"2026-02-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12932941/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147310924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-15eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooaf161
Andrew D Schechtman, Samantha M R Kling, Donn W Garvert, Jessica Y Lin, Tait Shanafelt, Marcy Winget, Christopher Sharp
Objectives: Electronic health record (EHR) order preference lists and order sets potentially improve efficiency but have limited utility in complex primary care settings. We assessed adoption, impact on ordering efficiency, and clinician perceptions of a comprehensive set of nested order panels (xOrders) for adult primary care.
Methods: In Phase 1 (gradual implementation), 404 xOrders were released (November 29, 2020-September 25, 2021). Beginning of Phase 2 (rapid implementation), 630 xOrders were released with an additional 253 xOrders added (September 26, 2021-June 24, 2023). Three outcomes captured adoption: xOrders used per week; number of clinician users per week; and percent of xOrders of all orders. Impact of xOrders on times in orders per encounter per clinician was evaluated with mixed effects interrupted time series. t-Tests evaluated differences between low, moderate, and high utilizers. A survey captured clinicians' perceptions in November 2022.
Results: xOrders were used 536 (SD, 245) times/week and by 57(15) clinicians/week in Phase 2. xOrders as a percent of all orders ranged from 0% to 31% across clinicians. Time spent in orders per encounter decreased by 14 ± 5 s (P =.01) from Phase 1 to 2 for high utilizers, decreased by 7(3) s (P=.05) for moderate utilizers, and increased by 1(3) s for low utilizers (P=.81); low and high utilizers were significantly different (P=.02). Most (77%) survey respondents agreed that xOrders improved ordering efficiency.
Discussion and conclusions: Despite yielding time savings and positive clinician feedback, the xOrder intervention showed limited adoption and impact, suggesting the need for expanded content and increased adoption to realize larger efficiency gains.
{"title":"Nested order panels for adult primary care modestly improves ordering efficiency among high utilizers.","authors":"Andrew D Schechtman, Samantha M R Kling, Donn W Garvert, Jessica Y Lin, Tait Shanafelt, Marcy Winget, Christopher Sharp","doi":"10.1093/jamiaopen/ooaf161","DOIUrl":"https://doi.org/10.1093/jamiaopen/ooaf161","url":null,"abstract":"<p><strong>Objectives: </strong>Electronic health record (EHR) order preference lists and order sets potentially improve efficiency but have limited utility in complex primary care settings. We assessed adoption, impact on ordering efficiency, and clinician perceptions of a comprehensive set of nested order panels (xOrders) for adult primary care.</p><p><strong>Methods: </strong>In <i>Phase 1</i> (gradual implementation), 404 xOrders were released (November 29, 2020-September 25, 2021). Beginning of <i>Phase 2</i> (rapid implementation), 630 xOrders were released with an additional 253 xOrders added (September 26, 2021-June 24, 2023). Three outcomes captured adoption: xOrders used per week; number of clinician users per week; and percent of xOrders of all orders. Impact of xOrders on times in orders per encounter per clinician was evaluated with mixed effects interrupted time series. <i>t</i>-Tests evaluated differences between low, moderate, and high utilizers. A survey captured clinicians' perceptions in November 2022.</p><p><strong>Results: </strong>xOrders were used 536 (SD, 245) times/week and by 57(15) clinicians/week in <i>Phase 2</i>. xOrders as a percent of all orders ranged from 0% to 31% across clinicians. Time spent in orders per encounter decreased by 14 ± 5 s (<i>P</i> =.01) from <i>Phase 1</i> to <i>2</i> for high utilizers, decreased by 7(3) s (<i>P</i>=.05) for moderate utilizers, and increased by 1(3) s for low utilizers (<i>P</i>=.81); low and high utilizers were significantly different (<i>P</i>=.02). Most (77%) survey respondents agreed that xOrders improved ordering efficiency.</p><p><strong>Discussion and conclusions: </strong>Despite yielding time savings and positive clinician feedback, the xOrder intervention showed limited adoption and impact, suggesting the need for expanded content and increased adoption to realize larger efficiency gains.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooaf161"},"PeriodicalIF":3.4,"publicationDate":"2026-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12907020/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146214520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-15eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooag011
Miranda Edmunds, Aditi Gupta, Inez Oh, Levi Kaster, Jonathan Sagel, Andrew Michelson, Yuyao Zhuge, Philip Payne, Ahmed Said
Objectives: To develop a multimodal pediatric critical care datamart supporting predictive modeling and decision support tool development, integrating high-resolution physiologic and clinical data and future clinical deployment.
Materials and methods: We developed a continuously expanding datamart integrating electronic health record data, high-resolution telemetry, and extracorporeal membrane oxygenation (ECMO)-domain datasets. The platform links static and longitudinal time-series variables with expert-curated neurologic outcomes for both ECMO and non-ECMO patients, enabling trajectory-based analyses.
Results: The datamart currently includes 25 762 pediatric patients, of whom 395 received ECMO support. The datamart captures granular longitudinal physiologic, laboratory, medication, and telemetry data suitable for dynamic predictive modeling.
Discussion: Existing ECMO prognostication tools rely on static variables and lack appropriate control cohorts. This datamart enables trajectory-based multimodal modeling, that reflects evolving physiology and neurologic outcomes.
Conclusion: This platform provides a scalable foundation for predictive modeling across pediatric critical care, beyond ECMO, to support precision decision-making and outcomes research.
{"title":"Multi-modal pediatric critical care datamart for extracorporeal support prediction and decision support.","authors":"Miranda Edmunds, Aditi Gupta, Inez Oh, Levi Kaster, Jonathan Sagel, Andrew Michelson, Yuyao Zhuge, Philip Payne, Ahmed Said","doi":"10.1093/jamiaopen/ooag011","DOIUrl":"https://doi.org/10.1093/jamiaopen/ooag011","url":null,"abstract":"<p><strong>Objectives: </strong>To develop a multimodal pediatric critical care datamart supporting predictive modeling and decision support tool development, integrating high-resolution physiologic and clinical data and future clinical deployment.</p><p><strong>Materials and methods: </strong>We developed a continuously expanding datamart integrating electronic health record data, high-resolution telemetry, and extracorporeal membrane oxygenation (ECMO)-domain datasets. The platform links static and longitudinal time-series variables with expert-curated neurologic outcomes for both ECMO and non-ECMO patients, enabling trajectory-based analyses.</p><p><strong>Results: </strong>The datamart currently includes 25 762 pediatric patients, of whom 395 received ECMO support. The datamart captures granular longitudinal physiologic, laboratory, medication, and telemetry data suitable for dynamic predictive modeling.</p><p><strong>Discussion: </strong>Existing ECMO prognostication tools rely on static variables and lack appropriate control cohorts. This datamart enables trajectory-based multimodal modeling, that reflects evolving physiology and neurologic outcomes.</p><p><strong>Conclusion: </strong>This platform provides a scalable foundation for predictive modeling across pediatric critical care, beyond ECMO, to support precision decision-making and outcomes research.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooag011"},"PeriodicalIF":3.4,"publicationDate":"2026-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12907016/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146214436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-15eCollection Date: 2026-02-01DOI: 10.1093/jamiaopen/ooaf170
Aman Saiju, Subiksha Umakanth, Anna Vaynrub, Romi Eli, Alissa Michel, Katherine D Crew, Rita Kukafka
Objectives: This study aims to develop a detailed understanding of provider and Information Technology (IT) operations staff experiences and attitudes regarding patients' ability to edit their data. This includes understanding barriers to developing a process to write back data into the electronic health record (EHR) as well as a concrete set of recommendations on incorporating patient-generated data into the EHR.
Materials and methods: RealRisks, our team's Fast Healthcare Interoperability Resources-compliant web-based patient decision aid, was utilized as an exemplar platform in which patients can access EHR data and review, correct, and contribute patient-derived data when specific elements are missing. An interview guide was developed and semi-structured interviews of 9 participants (physicians n = 4, IT operations staff n = 5) at Columbia University Irving Medical Center were carried out to understand the feasibility of writing back patient-entered edits into the EHR using the RealRisks decision aid.
Results: Providers and IT operations staff reported varied knowledge of how patients interact with their data but collectively stated a need for increasing EHR accuracy that prioritizes provider-patient communication. Participants supported a write-back process and had specific suggestions for implementation mechanisms (such as the option to upload test results when submitting changes).
Discussion: Providers and IT operations staff maintained that existing data management routes used for external data incorporation should be utilized, and that providers should screen edit requests to ensure EHR quality and accuracy.
Conclusion: While participants felt a write-back of patient-derived data would be helpful, future studies should directly assess nursing staff and advanced practice providers as well as patient perspectives to ensure equity and efficacy.
{"title":"Provider and information technology operations staff perspectives on the feasibility of writing patient-generated health data into the electronic health record.","authors":"Aman Saiju, Subiksha Umakanth, Anna Vaynrub, Romi Eli, Alissa Michel, Katherine D Crew, Rita Kukafka","doi":"10.1093/jamiaopen/ooaf170","DOIUrl":"https://doi.org/10.1093/jamiaopen/ooaf170","url":null,"abstract":"<p><strong>Objectives: </strong>This study aims to develop a detailed understanding of provider and Information Technology (IT) operations staff experiences and attitudes regarding patients' ability to edit their data. This includes understanding barriers to developing a process to write back data into the electronic health record (EHR) as well as a concrete set of recommendations on incorporating patient-generated data into the EHR.</p><p><strong>Materials and methods: </strong><i>RealRisks</i>, our team's Fast Healthcare Interoperability Resources-compliant web-based patient decision aid, was utilized as an exemplar platform in which patients can access EHR data and review, correct, and contribute patient-derived data when specific elements are missing. An interview guide was developed and semi-structured interviews of 9 participants (physicians <i>n</i> = 4, IT operations staff <i>n</i> = 5) at Columbia University Irving Medical Center were carried out to understand the feasibility of writing back patient-entered edits into the EHR using the <i>RealRisks</i> decision aid.</p><p><strong>Results: </strong>Providers and IT operations staff reported varied knowledge of how patients interact with their data but collectively stated a need for increasing EHR accuracy that prioritizes provider-patient communication. Participants supported a write-back process and had specific suggestions for implementation mechanisms (such as the option to upload test results when submitting changes).</p><p><strong>Discussion: </strong>Providers and IT operations staff maintained that existing data management routes used for external data incorporation should be utilized, and that providers should screen edit requests to ensure EHR quality and accuracy.</p><p><strong>Conclusion: </strong>While participants felt a write-back of patient-derived data would be helpful, future studies should directly assess nursing staff and advanced practice providers as well as patient perspectives to ensure equity and efficacy.</p>","PeriodicalId":36278,"journal":{"name":"JAMIA Open","volume":"9 1","pages":"ooaf170"},"PeriodicalIF":3.4,"publicationDate":"2026-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12907014/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146214459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}