Inventory management for emergency carts is one of the routine tasks in hospitals. It is highly desirable to simplify the workflow of the inventory task since healthcare staffs always work under high pressure and heavy workload. In this study, we exploit computer vision technology to develop an automated inventory management system for emergency carts. We have conducted a user study evaluate our proposed system. The subjects are 37 nurses from the internal medicine department and the pediatrics department of Chia-Yi Christian Hospital. The time spending on the inventory task before and after using our proposed system are compared. We also evaluate several state-of-the-art object detection algorithms (YOLOv4, YOLOv5, YOLOv7, and YOLOv5-OBB) for the task of automatic drug counting. We collect 500 images to train these object detection models and evaluate their accuracy. After using the proposed system, the time spent on the inventory task is reduced from 9.59 min to 5.32 min in average. The user study also indicates that the average workload score of the nurses is reduced from 3.81 to 2.70. For the evaluation of automatic drug counting, our proposed hybrid approach, which combines YOLOv5-OBB and YOLOv7, yields the highest accuracy. The resulting over detection rate and missed detection rate are 1.29% and 3.27%, respectively. In this study, we demonstrate that smartphone-based inventory management system using computer vision technology, specifically YOLO series object detectors, can effectively streamline inventory workflow and reduce the workload for healthcare personnel.
{"title":"Development of a Smartphone-Based Inventory Management System for Emergency Carts.","authors":"Chia-Hui Liu, Nian-Yin Wu, Che-Chia Liu, Wen-Yin Kuo, Tzu-Chia Lin, Wei-Yang Lin","doi":"10.1007/s10916-025-02268-y","DOIUrl":"https://doi.org/10.1007/s10916-025-02268-y","url":null,"abstract":"<p><p>Inventory management for emergency carts is one of the routine tasks in hospitals. It is highly desirable to simplify the workflow of the inventory task since healthcare staffs always work under high pressure and heavy workload. In this study, we exploit computer vision technology to develop an automated inventory management system for emergency carts. We have conducted a user study evaluate our proposed system. The subjects are 37 nurses from the internal medicine department and the pediatrics department of Chia-Yi Christian Hospital. The time spending on the inventory task before and after using our proposed system are compared. We also evaluate several state-of-the-art object detection algorithms (YOLOv4, YOLOv5, YOLOv7, and YOLOv5-OBB) for the task of automatic drug counting. We collect 500 images to train these object detection models and evaluate their accuracy. After using the proposed system, the time spent on the inventory task is reduced from 9.59 min to 5.32 min in average. The user study also indicates that the average workload score of the nurses is reduced from 3.81 to 2.70. For the evaluation of automatic drug counting, our proposed hybrid approach, which combines YOLOv5-OBB and YOLOv7, yields the highest accuracy. The resulting over detection rate and missed detection rate are 1.29% and 3.27%, respectively. In this study, we demonstrate that smartphone-based inventory management system using computer vision technology, specifically YOLO series object detectors, can effectively streamline inventory workflow and reduce the workload for healthcare personnel.</p>","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"133"},"PeriodicalIF":5.7,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145258369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-10DOI: 10.1007/s10916-025-02283-z
Menglin Tian, Shaolong Li, Wenyin Du, Sen Yang, Xiaohua Zhao, Hao Xiong, Hongxi Li, Mei Lu, Yunyan Ying, Jilei Zhang, Qiwei Liao, Dong Yang, Fuding Guo
The rapid evolution of large language models (LLMs) in the medical field, particularly in automating medical tasks and supporting diagnosis and treatment, has shown promising potential. However, their accuracy, comprehensiveness, and safety in managing complex cardiovascular diseases have not been systematically assessed. This study aims to evaluate and compare the diagnostic, therapeutic, and safety performance of two large language models-ChatGPT-4o and Kimi-in managing complex cardiovascular diseases, and to explore their potential for future clinical application. A total of 200 complex cardiovascular cases published in JACC: Case Reports between January 2020 and August 2024 were included. All cases were standardized and de-identified before being input into ChatGPT-4o and Kimi using identical prompts. Each model independently generated diagnostic, treatment, and long-term management plans. Three cardiovascular specialists independently evaluated the outputs in a blinded manner, scoring accuracy and comprehensiveness using Likert scales. Safety was assessed using a risk matrix analysis. Additionally, 50 cases were randomly selected for triangulation to compare model-generated recommendations with clinical guidelines. Statistical analysis was performed using the Wilcoxon signed-rank test with Benjamini-Hochberg correction for multiple comparisons. In preliminary diagnostic accuracy, the two models performed similarly (P = 0.663, r = 0.044), but ChatGPT-4o showed superior comprehensiveness (P < 0.001, r = 0.484). For treatment recommendations, ChatGPT-4o outperformed Kimi in both accuracy (P = 0.004, r = 0.321) and comprehensiveness (P < 0.001, r = 0.644). In long-term management, ChatGPT-4o demonstrated significant advantages in accuracy (P < 0.001, r = 0.717) and comprehensiveness (P < 0.001, r = 0.690). Safety assessment showed a lower proportion of high-risk outputs with ChatGPT-4o (1.5%) compared to Kimi (4.5%). LLMs, particularly ChatGPT-4o, exhibit significant promise in the diagnosis and treatment of complex cardiovascular diseases, showing superior accuracy, comprehensiveness, and safety compared to Kimi. Despite their high accuracy and safety, LLMs still require clinician oversight, especially in the formulation of personalized treatment plans and complex decision-making scenarios, to ensure their reliable integration into clinical practice.
大型语言模型(LLMs)在医疗领域的快速发展,特别是在自动化医疗任务和支持诊断和治疗方面,显示出了良好的潜力。然而,它们在治疗复杂心血管疾病中的准确性、全面性和安全性尚未得到系统评估。本研究旨在评估和比较chatgpt - 40和kimi这两种大型语言模型在治疗复杂心血管疾病中的诊断、治疗和安全性表现,并探讨其未来临床应用的潜力。纳入了JACC: 2020年1月至2024年8月期间发表的200例复杂心血管病例。在使用相同的提示输入chatgpt - 40和Kimi之前,所有病例都经过标准化和去识别。每个模型独立生成诊断、治疗和长期管理计划。三位心血管专家以盲法独立评估输出,使用李克特量表评分准确性和全面性。使用风险矩阵分析评估安全性。此外,随机选择50例病例进行三角测量,以比较模型生成的建议与临床指南。统计学分析采用Wilcoxon sign -rank检验,并采用Benjamini-Hochberg多重比较校正。在初步诊断准确性方面,两种模型表现相似(P = 0.663, r = 0.044),但chatgpt - 40表现出更强的全面性(P = 0.663, r = 0.044)
{"title":"Novel Insights into the Application of Large Language Models in the Diagnosis and Treatment of Complex Cardiovascular Diseases: A Comparative Study.","authors":"Menglin Tian, Shaolong Li, Wenyin Du, Sen Yang, Xiaohua Zhao, Hao Xiong, Hongxi Li, Mei Lu, Yunyan Ying, Jilei Zhang, Qiwei Liao, Dong Yang, Fuding Guo","doi":"10.1007/s10916-025-02283-z","DOIUrl":"https://doi.org/10.1007/s10916-025-02283-z","url":null,"abstract":"<p><p>The rapid evolution of large language models (LLMs) in the medical field, particularly in automating medical tasks and supporting diagnosis and treatment, has shown promising potential. However, their accuracy, comprehensiveness, and safety in managing complex cardiovascular diseases have not been systematically assessed. This study aims to evaluate and compare the diagnostic, therapeutic, and safety performance of two large language models-ChatGPT-4o and Kimi-in managing complex cardiovascular diseases, and to explore their potential for future clinical application. A total of 200 complex cardiovascular cases published in JACC: Case Reports between January 2020 and August 2024 were included. All cases were standardized and de-identified before being input into ChatGPT-4o and Kimi using identical prompts. Each model independently generated diagnostic, treatment, and long-term management plans. Three cardiovascular specialists independently evaluated the outputs in a blinded manner, scoring accuracy and comprehensiveness using Likert scales. Safety was assessed using a risk matrix analysis. Additionally, 50 cases were randomly selected for triangulation to compare model-generated recommendations with clinical guidelines. Statistical analysis was performed using the Wilcoxon signed-rank test with Benjamini-Hochberg correction for multiple comparisons. In preliminary diagnostic accuracy, the two models performed similarly (P = 0.663, r = 0.044), but ChatGPT-4o showed superior comprehensiveness (P < 0.001, r = 0.484). For treatment recommendations, ChatGPT-4o outperformed Kimi in both accuracy (P = 0.004, r = 0.321) and comprehensiveness (P < 0.001, r = 0.644). In long-term management, ChatGPT-4o demonstrated significant advantages in accuracy (P < 0.001, r = 0.717) and comprehensiveness (P < 0.001, r = 0.690). Safety assessment showed a lower proportion of high-risk outputs with ChatGPT-4o (1.5%) compared to Kimi (4.5%). LLMs, particularly ChatGPT-4o, exhibit significant promise in the diagnosis and treatment of complex cardiovascular diseases, showing superior accuracy, comprehensiveness, and safety compared to Kimi. Despite their high accuracy and safety, LLMs still require clinician oversight, especially in the formulation of personalized treatment plans and complex decision-making scenarios, to ensure their reliable integration into clinical practice.</p>","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"135"},"PeriodicalIF":5.7,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145274814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-10DOI: 10.1007/s10916-025-02281-1
Johnathan Loh, Yong Jin Tan, Kimberly Joseph, Roshan Sabeeha, Robby M Alpas, Lesmer S Beltran, Bryston C Torres, Fook Sang Wong, Abigail Pc Wong, Jerome A Poblete, Larry R Natividad, Cai Chao, Chee Yuen Lu, Kok Wei Aik, Hwee Thiang Tan, Xue Fen Teng, Jimin Liu, Felix Yj Keng, Angela S Koh
Background: Nuclear imaging is the cornerstone of clinical practice across many disciplines. Few innovations in imaging have addressed occupational health of radiographers exposed to radiation in their daily work. In this proof-of-concept study, we hypothesized that the use of autonomous robots in a high-volume Myocardial Perfusion Imaging (MPI) clinic to help reduce radiation risk for radiographers.
Methods: After initial assessment of the radiographer's workflow, an autonomous robot was set up to navigate and deliver radiotracer doses between the radio pharmacy and the injection room. 8 radiographers were briefed on robot usage and allowed to use the robot for their daily operations. Radiation exposure was measured as part of the regular bi-monthly Thermoluminescent Dosimeter (TLD) dose tracking. Radiation exposure before and after robot implementation were compared to assess whether the implementation of the autonomous robot significantly reduced radiation exposure for the radiographers.
Results: We observed a significant reduction in mean bi-monthly radiation exposure following the implementation of the autonomous robot. The mean radiation dose decreased from 0.67 mSv (95% CI: 0.60-0.74) pre-implementation to 0.49 mSv 95% (CI: 0.44-0.54) post-implementation, corresponding to a relative reduction of approximately 27%. A Bayesian independent t-test revealed strong evidence for this reduction, with Bayes Factors (BF10) of 17.04 for skin dose and 17.76 for whole-body dose, supporting the hypothesis that the autonomous robot effectively reduced radiation exposure among radiographers.
Conclusion: In this study, we provided a proof-of-concept on the use of autonomous robots in a MPI clinic as an additional tool to help radiographers manage radiation risk in their work. The innovations in technologies could expand the strategies available for managing occupational radiation risks in alignment with as low as reasonably achievable (ALARA) principles. Future work on scalability across diverse clinical and operational contexts would be next steps.
{"title":"Evaluation of an Autonomous Robotic System for Reducing Radiation Risk in a Real-World Cardiac Imaging Laboratory.","authors":"Johnathan Loh, Yong Jin Tan, Kimberly Joseph, Roshan Sabeeha, Robby M Alpas, Lesmer S Beltran, Bryston C Torres, Fook Sang Wong, Abigail Pc Wong, Jerome A Poblete, Larry R Natividad, Cai Chao, Chee Yuen Lu, Kok Wei Aik, Hwee Thiang Tan, Xue Fen Teng, Jimin Liu, Felix Yj Keng, Angela S Koh","doi":"10.1007/s10916-025-02281-1","DOIUrl":"https://doi.org/10.1007/s10916-025-02281-1","url":null,"abstract":"<p><strong>Background: </strong>Nuclear imaging is the cornerstone of clinical practice across many disciplines. Few innovations in imaging have addressed occupational health of radiographers exposed to radiation in their daily work. In this proof-of-concept study, we hypothesized that the use of autonomous robots in a high-volume Myocardial Perfusion Imaging (MPI) clinic to help reduce radiation risk for radiographers.</p><p><strong>Methods: </strong>After initial assessment of the radiographer's workflow, an autonomous robot was set up to navigate and deliver radiotracer doses between the radio pharmacy and the injection room. 8 radiographers were briefed on robot usage and allowed to use the robot for their daily operations. Radiation exposure was measured as part of the regular bi-monthly Thermoluminescent Dosimeter (TLD) dose tracking. Radiation exposure before and after robot implementation were compared to assess whether the implementation of the autonomous robot significantly reduced radiation exposure for the radiographers.</p><p><strong>Results: </strong>We observed a significant reduction in mean bi-monthly radiation exposure following the implementation of the autonomous robot. The mean radiation dose decreased from 0.67 mSv (95% CI: 0.60-0.74) pre-implementation to 0.49 mSv 95% (CI: 0.44-0.54) post-implementation, corresponding to a relative reduction of approximately 27%. A Bayesian independent t-test revealed strong evidence for this reduction, with Bayes Factors (BF10) of 17.04 for skin dose and 17.76 for whole-body dose, supporting the hypothesis that the autonomous robot effectively reduced radiation exposure among radiographers.</p><p><strong>Conclusion: </strong>In this study, we provided a proof-of-concept on the use of autonomous robots in a MPI clinic as an additional tool to help radiographers manage radiation risk in their work. The innovations in technologies could expand the strategies available for managing occupational radiation risks in alignment with as low as reasonably achievable (ALARA) principles. Future work on scalability across diverse clinical and operational contexts would be next steps.</p>","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"136"},"PeriodicalIF":5.7,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145274838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-10DOI: 10.1007/s10916-025-02279-9
Nathan Orwig, Asif Padiyath
{"title":"Title: The 90th Percentile Dilemma: Time Metrics and Real-World Feasibility in Preoperative Regional Anesthesia: A Commentary.","authors":"Nathan Orwig, Asif Padiyath","doi":"10.1007/s10916-025-02279-9","DOIUrl":"10.1007/s10916-025-02279-9","url":null,"abstract":"","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"134"},"PeriodicalIF":5.7,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145258298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-08DOI: 10.1007/s10916-025-02264-2
Shuai Ming, Xi Yao, Qingge Guo, Dandan Chen, Xiaohong Guo, Kunpeng Xie, Bo Lei
Background: DeepSeek-R1, an open-source reasoning large language model (LLM) clinically deployed in Chinese hospitals, still lacks validation in ophthalmology.
Aims: To compare DeepSeek-R1 against OpenAI's o1 and upgraded o3 models in diagnostic accuracy and reasoning capability across diverse ophthalmic conditions.
Methods: We evaluated 98 standardized case vignettes covering13 ophthalmic sub-specialties, each supplied with an expert-validated diagnostic hierarchy, differential list, and reasoning chain. Model performance was assessed with a diagnosis matrix focused on final-diagnosis (FDx) accuracy; incorrect outputs were resubmitted with key diagnostic clues (reasoning-augmented, RA prompt) to test self-correction. Reasoning capacity was quantified by the number/score of diagnostic clues retrieved per case across 13 predefined domains.
Results: DeepSeek-R1 achieved an 87.8% FDx accuracy, comparable to o3 (91.8%, P = .34) and higher than o1 (58.2%, P < .001). Similar trends were observed for others accuracy (global P < .001). Agreement was moderate-high between R1 and o3 (κ = 0.42-1.00), but slight with o1 (κ = 0.12-0.32). R1 and o3 identified more diagnostic clues than o1 (median count = 4 vs. 3, median score = 100 vs. 80; P < .001). RA prompts corrected 50.0%, 62.5% and 41.5% of FDx errors for R1, o3, and o1, raising FDx accuracy to 93.9%, 96.9%, and 80.6% respectively.
Conclusions: DeepSeek-R1 matched o3 and outperformed o1 in diagnostic accuracy and reasoning, retrieving nearly all expert-defined clues. Its open-source nature, low cost and strong performance support its use as a practical aid for ophthalmic decision-making.
背景:DeepSeek-R1是中国医院临床部署的开源推理大语言模型(LLM),但仍缺乏在眼科领域的验证。目的:比较DeepSeek-R1与OpenAI的o1和升级后的o3模型在不同眼科条件下的诊断准确性和推理能力。方法:我们评估了涵盖13个眼科亚专科的98个标准化病例,每个病例都提供了专家验证的诊断层次、差异列表和推理链。用诊断矩阵评估模型的性能,以最终诊断(FDx)的准确性为重点;不正确的输出与关键诊断线索(推理增强,RA提示)一起重新提交,以测试自我纠正。推理能力通过在13个预定义域中检索到的每个病例的诊断线索的数量/分数来量化。结果:DeepSeek-R1的FDx准确率为87.8%,与o3 (91.8%, P =。结论:DeepSeek-R1在诊断准确性和推理方面与o3相匹配,并且在检索几乎所有专家定义的线索方面优于o1。它的开源性质、低成本和强大的性能支持它作为眼科决策的实用辅助工具。
{"title":"Evaluation of DeepSeek-R1 for Ophthalmic Diagnosis and Reasoning: A Comparison with OpenAI o1 and o3.","authors":"Shuai Ming, Xi Yao, Qingge Guo, Dandan Chen, Xiaohong Guo, Kunpeng Xie, Bo Lei","doi":"10.1007/s10916-025-02264-2","DOIUrl":"https://doi.org/10.1007/s10916-025-02264-2","url":null,"abstract":"<p><strong>Background: </strong>DeepSeek-R1, an open-source reasoning large language model (LLM) clinically deployed in Chinese hospitals, still lacks validation in ophthalmology.</p><p><strong>Aims: </strong>To compare DeepSeek-R1 against OpenAI's o1 and upgraded o3 models in diagnostic accuracy and reasoning capability across diverse ophthalmic conditions.</p><p><strong>Methods: </strong>We evaluated 98 standardized case vignettes covering13 ophthalmic sub-specialties, each supplied with an expert-validated diagnostic hierarchy, differential list, and reasoning chain. Model performance was assessed with a diagnosis matrix focused on final-diagnosis (FDx) accuracy; incorrect outputs were resubmitted with key diagnostic clues (reasoning-augmented, RA prompt) to test self-correction. Reasoning capacity was quantified by the number/score of diagnostic clues retrieved per case across 13 predefined domains.</p><p><strong>Results: </strong>DeepSeek-R1 achieved an 87.8% FDx accuracy, comparable to o3 (91.8%, P = .34) and higher than o1 (58.2%, P < .001). Similar trends were observed for others accuracy (global P < .001). Agreement was moderate-high between R1 and o3 (κ = 0.42-1.00), but slight with o1 (κ = 0.12-0.32). R1 and o3 identified more diagnostic clues than o1 (median count = 4 vs. 3, median score = 100 vs. 80; P < .001). RA prompts corrected 50.0%, 62.5% and 41.5% of FDx errors for R1, o3, and o1, raising FDx accuracy to 93.9%, 96.9%, and 80.6% respectively.</p><p><strong>Conclusions: </strong>DeepSeek-R1 matched o3 and outperformed o1 in diagnostic accuracy and reasoning, retrieving nearly all expert-defined clues. Its open-source nature, low cost and strong performance support its use as a practical aid for ophthalmic decision-making.</p>","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"130"},"PeriodicalIF":5.7,"publicationDate":"2025-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145251448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-08DOI: 10.1007/s10916-025-02271-3
Tianyuan Gan, Chongan Zhang, Peng Wang, Xiao Liang, Xuesong Ye
Accurate polyp detection is essential for the early diagnosis and effective treatment of colorectal cancer (CRC). However, colonoscopy videos in real-world clinical settings present significant challenges, often causing existing algorithms to fail. Compared to single images, videos contain richer temporal and contextual information, making them valuable for developing deep-learning-based detection systems. To address these challenges, we propose an end-to-end Two-Stream Polyp Detection Transformer (TS-PDTR) network. First, our framework uses a two-stream feature extraction network to capture both spatial and temporal features from the RGB frames and optical flow. Then, the proposed Detail-Aware Convolution (DAConv) module enhances fine-grained contextual information in low-level features. Following this, the Detail-Guided Attention (DGA) module generates channel-specific Spatial Attention Maps (SAMs) to refine deep feature maps, improving the model's sensitivity to small and camouflaged polyps. Finally, a Flow Fusion Encoder (FFE) module combines temporal cues from optical flow to increase robustness against poor single-frame image quality. Experiments on three benchmark video colonoscopy datasets show that TS-PDTR consistently outperforms previous state-of-the-art image- and video-based polyp detection methods. Notably, our model achieves a mean Average Precision (mAP) of 33.2 on the most challenging LDPolypVideo dataset. It also improves the mAP to 64.0 and 55.6 on the SUN Colonoscopy Video Database and CVC-VideoClinicDB, respectively. In summary, TS-PDTR is a promising video-based polyp detection method with strong potential for further development and real-world clinical application.
{"title":"Revisiting Challenges in Real-world Video Colonoscopy using End-to-End Two Stream Polyp Detection Transformer (TS-PDTR).","authors":"Tianyuan Gan, Chongan Zhang, Peng Wang, Xiao Liang, Xuesong Ye","doi":"10.1007/s10916-025-02271-3","DOIUrl":"https://doi.org/10.1007/s10916-025-02271-3","url":null,"abstract":"<p><p>Accurate polyp detection is essential for the early diagnosis and effective treatment of colorectal cancer (CRC). However, colonoscopy videos in real-world clinical settings present significant challenges, often causing existing algorithms to fail. Compared to single images, videos contain richer temporal and contextual information, making them valuable for developing deep-learning-based detection systems. To address these challenges, we propose an end-to-end Two-Stream Polyp Detection Transformer (TS-PDTR) network. First, our framework uses a two-stream feature extraction network to capture both spatial and temporal features from the RGB frames and optical flow. Then, the proposed Detail-Aware Convolution (DAConv) module enhances fine-grained contextual information in low-level features. Following this, the Detail-Guided Attention (DGA) module generates channel-specific Spatial Attention Maps (SAMs) to refine deep feature maps, improving the model's sensitivity to small and camouflaged polyps. Finally, a Flow Fusion Encoder (FFE) module combines temporal cues from optical flow to increase robustness against poor single-frame image quality. Experiments on three benchmark video colonoscopy datasets show that TS-PDTR consistently outperforms previous state-of-the-art image- and video-based polyp detection methods. Notably, our model achieves a mean Average Precision (mAP) of 33.2 on the most challenging LDPolypVideo dataset. It also improves the mAP to 64.0 and 55.6 on the SUN Colonoscopy Video Database and CVC-VideoClinicDB, respectively. In summary, TS-PDTR is a promising video-based polyp detection method with strong potential for further development and real-world clinical application.</p>","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"129"},"PeriodicalIF":5.7,"publicationDate":"2025-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145251451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-07DOI: 10.1007/s10916-025-02276-y
Christopher Reagen, Katelyn M Brown, Thomas Hocker
Artificial intelligence (AI), specifically large language models (LLM), have gained significant popularity over the last decade with increased performance and expanding applications. AI could improve the quality of patient care in medicine but hidden biases introduced during training could be harmful. This work utilizes GPT-4o-mini to generate patient communications based on systematically generated, synthetic patient data that would be commonly available in a patient's medical record. To evaluate the AI generated communications for disparities, GPT-4o-mini was used to score the generated communications on empathy, encouragement, accuracy, clarity, professionalism, and respect. Disparities in scores associated with specific components of a patient's history were used to detect potential biases. A patient's sex and religious preference were found to have a statistically significant impact on scores. However, further work is needed to evaluate a wider collection of LLMs utilizing more specific and human validated scoring criteria. Overall, this work proposes a novel method of evaluating bias in LLMs by creating synthetic patient histories to formulate AI generated communications and score them with opportunities for further investigation.
{"title":"Invisible Bias in GPT-4o-mini: Detecting Disparities in AI-Generated Patient Messaging.","authors":"Christopher Reagen, Katelyn M Brown, Thomas Hocker","doi":"10.1007/s10916-025-02276-y","DOIUrl":"https://doi.org/10.1007/s10916-025-02276-y","url":null,"abstract":"<p><p>Artificial intelligence (AI), specifically large language models (LLM), have gained significant popularity over the last decade with increased performance and expanding applications. AI could improve the quality of patient care in medicine but hidden biases introduced during training could be harmful. This work utilizes GPT-4o-mini to generate patient communications based on systematically generated, synthetic patient data that would be commonly available in a patient's medical record. To evaluate the AI generated communications for disparities, GPT-4o-mini was used to score the generated communications on empathy, encouragement, accuracy, clarity, professionalism, and respect. Disparities in scores associated with specific components of a patient's history were used to detect potential biases. A patient's sex and religious preference were found to have a statistically significant impact on scores. However, further work is needed to evaluate a wider collection of LLMs utilizing more specific and human validated scoring criteria. Overall, this work proposes a novel method of evaluating bias in LLMs by creating synthetic patient histories to formulate AI generated communications and score them with opportunities for further investigation.</p>","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"127"},"PeriodicalIF":5.7,"publicationDate":"2025-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145238838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-07DOI: 10.1007/s10916-025-02265-1
Caleb Keng, Anthony DiGiorgio, Jesse M Ehrenfeld, Joseph Spear, Brian J Miller
Healthcare delivery systems face mounting administrative complexity that contributes to clinician burnout, medical errors, and reduced access to care for patients. This editorial explores how automation and artificial intelligence (AI) can address key operational inefficiencies-specifically in prior authorization, quality metric reporting, and clinical documentation-by leveraging informatics-driven solutions. We examine the current landscape, quantify the impact of administrative burden, and propose informatics strategies to realign healthcare delivery around patient-centered, efficient care.
{"title":"Unburdening Patients and Clinicians Through Automation and Artificial Intelligence: Informatics Strategies for Reducing Administrative Burden.","authors":"Caleb Keng, Anthony DiGiorgio, Jesse M Ehrenfeld, Joseph Spear, Brian J Miller","doi":"10.1007/s10916-025-02265-1","DOIUrl":"10.1007/s10916-025-02265-1","url":null,"abstract":"<p><p>Healthcare delivery systems face mounting administrative complexity that contributes to clinician burnout, medical errors, and reduced access to care for patients. This editorial explores how automation and artificial intelligence (AI) can address key operational inefficiencies-specifically in prior authorization, quality metric reporting, and clinical documentation-by leveraging informatics-driven solutions. We examine the current landscape, quantify the impact of administrative burden, and propose informatics strategies to realign healthcare delivery around patient-centered, efficient care.</p>","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"128"},"PeriodicalIF":5.7,"publicationDate":"2025-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12504360/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145238864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-07DOI: 10.1007/s10916-025-02270-4
Bruno Caracci, Will Dixon, Rilong Zhang, German Serrano, Spencer Vecile, Andrew Goodwin, Robert Greer, Clyde Matava, Alex Mariakakis, Asad Siddiqui
Normative ranges for vital signs under general anesthesia are well established for healthy pediatric patients, but the influence of The American Society of Anesthesiologists Physical Status (ASA-PS) classification on these normative ranges remains unexplored. The purpose of this study is to develop age-based normative ranges for heart rate (HR) and blood pressure (BP) in patients undergoing general anesthesia for noncardiac surgery in our institution and assess differences by ASA-PS classification. This is a retrospective observational single-center study. We reviewed all anesthetic records from the Hospital for Sick Children, Canada between March 1st and December 31st, 2023. We extracted physiological data from our in-house high-resolution physiological data repository (AtriumDB) to develop normative ranges for physiological parameters and compared them according to ASA-PS classification. We developed age-based normative ranges for BP and HR. We found significant differences between ASA-PS groups, most notably between ASA-PS 1 and 5. We found a statistically significant difference between ASA-PS 1-2 and 3-5 across all physiological parameters. This study validates existing pediatric anesthesia reference ranges while demonstrating the feasibility of incorporating patients across the spectrum of ASA-PS. Further multicenter studies are needed to generalize these findings.
{"title":"Development of Normative Ranges for Vital Signs and Differentiation by American Society of Anesthesiologists Physical Status Category: A Retrospective Observational Study.","authors":"Bruno Caracci, Will Dixon, Rilong Zhang, German Serrano, Spencer Vecile, Andrew Goodwin, Robert Greer, Clyde Matava, Alex Mariakakis, Asad Siddiqui","doi":"10.1007/s10916-025-02270-4","DOIUrl":"10.1007/s10916-025-02270-4","url":null,"abstract":"<p><p>Normative ranges for vital signs under general anesthesia are well established for healthy pediatric patients, but the influence of The American Society of Anesthesiologists Physical Status (ASA-PS) classification on these normative ranges remains unexplored. The purpose of this study is to develop age-based normative ranges for heart rate (HR) and blood pressure (BP) in patients undergoing general anesthesia for noncardiac surgery in our institution and assess differences by ASA-PS classification. This is a retrospective observational single-center study. We reviewed all anesthetic records from the Hospital for Sick Children, Canada between March 1st and December 31st, 2023. We extracted physiological data from our in-house high-resolution physiological data repository (AtriumDB) to develop normative ranges for physiological parameters and compared them according to ASA-PS classification. We developed age-based normative ranges for BP and HR. We found significant differences between ASA-PS groups, most notably between ASA-PS 1 and 5. We found a statistically significant difference between ASA-PS 1-2 and 3-5 across all physiological parameters. This study validates existing pediatric anesthesia reference ranges while demonstrating the feasibility of incorporating patients across the spectrum of ASA-PS. Further multicenter studies are needed to generalize these findings.</p>","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"126"},"PeriodicalIF":5.7,"publicationDate":"2025-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145238849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-04DOI: 10.1007/s10916-025-02255-3
Alina Vozna, Andrea Monaldini, Stefania Costantini
This paper presents a trust-aware architecture for personalized digital health that combines user modeling, symbolic reasoning, and adaptive trust mechanisms. The proposed system uses Blueprint Personas to capture detailed patient profiles, including clinical, behavioral, and emotional traits. These profiles guide an intelligent agent that interacts with patients and healthcare professionals to provide context-sensitive support. Personalization is achieved through an ontology-based reasoning layer that interprets user needs and integrates real-time data from electronic health records, wearable devices, and environmental sources. To promote transparency and foster long-term user engagement, the system includes a formal trust modeling component based on a Reference Ontology of Trust (ROT), allowing the system to flexibly tailor communication strategies in response to user feedback and evolving trust levels. A simulated scenario involving a patient with chronic obstructive pulmonary disease demonstrates how the system delivers proactive and personalized healthcare interventions, such as medication reminders and air quality alerts. While the architecture is modular and designed for scalability, it has not yet been deployed in real-world clinical settings. Empirical validation and integration with clinical platforms remain part of future work. Nevertheless, this ongoing work contributes to the development of explainable and ethically aligned AI systems that enhance autonomy, accessibility, and trust in digital health environments through explainable reasoning.
本文提出了一种个性化数字健康的信任感知架构,该架构结合了用户建模、符号推理和自适应信任机制。该系统使用Blueprint Personas来获取详细的患者资料,包括临床、行为和情感特征。这些配置文件指导智能代理与患者和医疗保健专业人员进行交互,以提供上下文敏感的支持。个性化是通过基于本体的推理层实现的,该推理层解释用户需求并集成来自电子健康记录、可穿戴设备和环境来源的实时数据。为了提高透明度和促进长期用户参与,该系统包括一个基于信任参考本体(Reference Ontology of trust, ROT)的正式信任建模组件,使系统能够根据用户反馈和不断变化的信任水平灵活定制沟通策略。一个涉及慢性阻塞性肺病患者的模拟场景演示了该系统如何提供主动和个性化的医疗干预措施,例如药物提醒和空气质量警报。虽然该体系结构是模块化的,并为可扩展性而设计,但它尚未在现实世界的临床环境中部署。实证验证和与临床平台的整合仍然是未来工作的一部分。尽管如此,这项正在进行的工作有助于开发可解释和符合伦理的人工智能系统,通过可解释的推理增强数字卫生环境中的自主性、可及性和信任。
{"title":"A Trust-Aware Architecture for Personalized Digital Health: Integrating Blueprint Personas and Ontology-Based Reasoning.","authors":"Alina Vozna, Andrea Monaldini, Stefania Costantini","doi":"10.1007/s10916-025-02255-3","DOIUrl":"10.1007/s10916-025-02255-3","url":null,"abstract":"<p><p>This paper presents a trust-aware architecture for personalized digital health that combines user modeling, symbolic reasoning, and adaptive trust mechanisms. The proposed system uses Blueprint Personas to capture detailed patient profiles, including clinical, behavioral, and emotional traits. These profiles guide an intelligent agent that interacts with patients and healthcare professionals to provide context-sensitive support. Personalization is achieved through an ontology-based reasoning layer that interprets user needs and integrates real-time data from electronic health records, wearable devices, and environmental sources. To promote transparency and foster long-term user engagement, the system includes a formal trust modeling component based on a Reference Ontology of Trust (ROT), allowing the system to flexibly tailor communication strategies in response to user feedback and evolving trust levels. A simulated scenario involving a patient with chronic obstructive pulmonary disease demonstrates how the system delivers proactive and personalized healthcare interventions, such as medication reminders and air quality alerts. While the architecture is modular and designed for scalability, it has not yet been deployed in real-world clinical settings. Empirical validation and integration with clinical platforms remain part of future work. Nevertheless, this ongoing work contributes to the development of explainable and ethically aligned AI systems that enhance autonomy, accessibility, and trust in digital health environments through explainable reasoning.</p>","PeriodicalId":16338,"journal":{"name":"Journal of Medical Systems","volume":"49 1","pages":"125"},"PeriodicalIF":5.7,"publicationDate":"2025-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12496277/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145225557","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}