首页 > 最新文献

JMIR Medical Education最新文献

英文 中文
The Effect of a Traditional Chinese Medicine Course on Western Medicine Students' Attitudes Toward Traditional Chinese Medicine: Self-Controlled Pre-Post Questionnaire Study. 中医课程对西医学生中医态度的影响:自我控制的前后问卷研究。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-16 DOI: 10.2196/55972
Haoyu He, Hanjin Cui, Li Guo, Yanhui Liao

Background: Traditional Chinese medicine (TCM) hasbeen widely used to treat various diseases in China for thousands of years and has shown satisfactory effectiveness. However, many surveys found that TCM receives little recognition from Western medicine (WM) physicians and students. At present, TCM is offered as a compulsory course for WM students in WM schools.

Objective: This study aimed to investigate whether TCM courses can affect the WM students' attitude toward TCM.

Methods: WM students from Xiangya Medical School were invited to completeaweb-based questionnaire before and immediately after a TCM course. Their attitude toward TCM and treatment preferences for different kinds of diseases were tested. The Attitude Scale of TCM (ASTCM) was used. The main part of the ASTCM was designed to measure the attitude of medical students towardTCM. It consisted of 18 items, divided into cognitive dimension (5 terms), emotional dimension (8 terms), and behavioral tendencyfactor (5 terms).

Results: Finally, the results of 118 five-year program (FYP) and 36 eight-year program (EYP) students were included. For FYP students, there was a significant increase in the total mean score (66.42, SD 7.66 vs 71.43, SD 7.38;P<.001) of ASTCM after the TCM course. Significant increases in mean scores of the 3 factors of attitude (cognition: 21.64, SD 2.08 vs 22.90, SD 1.94; affection: 25.21, SD 4.39 vs 27.96, SD 4.4; and behavioral tendency: 19.577, SD 3.02 vs 20.58, SD 2.76; P<.001)were also observed. Except for the score of behavioral tendency (17.50, SD 3.54 vs 18.78, SD 3.22; P=.02), a significant increase was not detected in total score, cognition, and affection in EPY students (total score: mean 60.36, SD 10.53 vs mean 62.92, SD 10.05; cognition: mean 20.50, SD 2.73 vs mean 20.69, SD 2.73; and affection: mean 22.36, SD 6.32 vs mean 23.44, SD 5.84; all P>.05). The treatment preference of FYP students in acute (P=.02), chronic (P=.003), and physical diseases (P=.02) showed remarkable change. A major change was also detected in internal diseases (P=.02), surgical diseases (perioperative period; P=.01), and mental illnesses (P=.02) in EYP students. This change mainly appeared as a decline in WM preference and an increase in TCM and WM preference.

Conclusions: The study showed that earlier exposure to the TCM course increased the positive attitude toward TCM in students majoring in WM. The results provide some suggestions for arraging TCM courses in WM schools.

背景:几千年来,中医在中国被广泛用于治疗各种疾病,并显示出令人满意的效果。然而,许多调查发现,中医很少得到西医医生和学生的认可。目前,中医已成为中医院校中西医学生的必修课程。目的:本研究旨在探讨中医课程是否会影响中医学生对中医的态度。方法:邀请湘雅医学院中医专业学生在中医课程前后分别填写网络问卷。调查了他们对中医的态度和对不同疾病的治疗偏好。采用中医态度量表(ASTCM)。ASTCM的主体部分旨在测量医学生对中医的态度。量表由18个条目组成,分为认知维度(5项)、情感维度(8项)和行为倾向因素(5项)。结果:最终纳入118名五年制(FYP)学生和36名八年制(EYP)学生的结果。五年期学生的总平均得分显著提高(66.42,SD 7.66 vs 71.43, SD 7.38;P.05)。五年级学生对急性(P= 0.02)、慢性(P= 0.003)和躯体疾病(P= 0.02)的治疗偏好有显著变化。EYP学生的内科疾病(P= 0.02)、外科疾病(围手术期P= 0.01)和精神疾病(P= 0.02)也发生了重大变化。这种变化主要表现为对中西医结合的偏好下降,对中西医结合的偏好增加。结论:研究表明,早期接触中医课程增加了中医专业学生对中医的积极态度。研究结果为中医院校中医课程的安排提供了一些建议。
{"title":"The Effect of a Traditional Chinese Medicine Course on Western Medicine Students' Attitudes Toward Traditional Chinese Medicine: Self-Controlled Pre-Post Questionnaire Study.","authors":"Haoyu He, Hanjin Cui, Li Guo, Yanhui Liao","doi":"10.2196/55972","DOIUrl":"https://doi.org/10.2196/55972","url":null,"abstract":"<p><strong>Background: </strong>Traditional Chinese medicine (TCM) hasbeen widely used to treat various diseases in China for thousands of years and has shown satisfactory effectiveness. However, many surveys found that TCM receives little recognition from Western medicine (WM) physicians and students. At present, TCM is offered as a compulsory course for WM students in WM schools.</p><p><strong>Objective: </strong>This study aimed to investigate whether TCM courses can affect the WM students' attitude toward TCM.</p><p><strong>Methods: </strong>WM students from Xiangya Medical School were invited to completeaweb-based questionnaire before and immediately after a TCM course. Their attitude toward TCM and treatment preferences for different kinds of diseases were tested. The Attitude Scale of TCM (ASTCM) was used. The main part of the ASTCM was designed to measure the attitude of medical students towardTCM. It consisted of 18 items, divided into cognitive dimension (5 terms), emotional dimension (8 terms), and behavioral tendencyfactor (5 terms).</p><p><strong>Results: </strong>Finally, the results of 118 five-year program (FYP) and 36 eight-year program (EYP) students were included. For FYP students, there was a significant increase in the total mean score (66.42, SD 7.66 vs 71.43, SD 7.38;P<.001) of ASTCM after the TCM course. Significant increases in mean scores of the 3 factors of attitude (cognition: 21.64, SD 2.08 vs 22.90, SD 1.94; affection: 25.21, SD 4.39 vs 27.96, SD 4.4; and behavioral tendency: 19.577, SD 3.02 vs 20.58, SD 2.76; P<.001)were also observed. Except for the score of behavioral tendency (17.50, SD 3.54 vs 18.78, SD 3.22; P=.02), a significant increase was not detected in total score, cognition, and affection in EPY students (total score: mean 60.36, SD 10.53 vs mean 62.92, SD 10.05; cognition: mean 20.50, SD 2.73 vs mean 20.69, SD 2.73; and affection: mean 22.36, SD 6.32 vs mean 23.44, SD 5.84; all P>.05). The treatment preference of FYP students in acute (P=.02), chronic (P=.003), and physical diseases (P=.02) showed remarkable change. A major change was also detected in internal diseases (P=.02), surgical diseases (perioperative period; P=.01), and mental illnesses (P=.02) in EYP students. This change mainly appeared as a decline in WM preference and an increase in TCM and WM preference.</p><p><strong>Conclusions: </strong>The study showed that earlier exposure to the TCM course increased the positive attitude toward TCM in students majoring in WM. The results provide some suggestions for arraging TCM courses in WM schools.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e55972"},"PeriodicalIF":3.2,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145991326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data Science Education for Residents, Researchers, and Students in Psychiatry and Psychology: Program Development and Evaluation Study. 精神病学和心理学的居民、研究人员和学生的数据科学教育:项目开发和评估研究。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-16 DOI: 10.2196/75125
Hayoung K Donnelly, David Mandell, Sy Hwang, Emily Schriver, Ugurcan Vurgun, Graydon Neill, Esha Patel, Megan E Reilly, Michael Steinberg, Amber Calloway, Robert Gallop, Maria A Oquendo, Gregory K Brown, Danielle L Mowery
<p><strong>Background: </strong>The use of artificial intelligence (AI) to analyze health care data has become common in behavioral health sciences. However, the lack of training opportunities for mental health professionals limits clinicians' ability to adopt AI in clinical settings. AI education is essential for trainees, equipping them with the literacy needed to implement AI tools in practice, collaborate effectively with data scientists, and develop skills as interdisciplinary researchers with computing skills.</p><p><strong>Objective: </strong>As part of the Penn Innovation in Suicide Prevention Implementation Research Center, we developed, implemented, and evaluated a virtual workshop to educate psychiatry and psychology trainees on using AI for suicide prevention research.</p><p><strong>Methods: </strong>The workshop introduced trainees to natural language processing (NLP) concepts and Python coding skills using Jupyter notebooks within a secure Microsoft Azure Databricks cloud computing and analytics environment. We designed a 3-hour workshop that covered 4 key NLP topics: data characterization, data standardization, concept extraction, and statistical analysis. To demonstrate real-world applications, we processed chief complaints from electronic health records to compare the prevalence of suicide-related encounters across populations by race, ethnicity, and age. Training materials were developed based on standard NLP techniques and domain-specific tasks, such as preprocessing psychiatry-related acronyms. Two researchers drafted and demonstrated the code, incorporating feedback from the Methods Core of the Innovation in Suicide Prevention Implementation Research to refine the materials. To evaluate the effectiveness of the workshop, we used the Kirkpatrick program evaluation model, focusing on participants' reactions (level 1) and learning outcomes (level 2). Confidence changes in knowledge and skills before and after the workshop were assessed using paired t tests, and open-ended questions were included to gather feedback for future improvements.</p><p><strong>Results: </strong>A total of 10 trainees participated in the workshop virtually, including residents, postdoctoral researchers, and graduate students from the psychiatry and psychology departments. The participants found the workshop helpful (mean 3.17 on a scale of 1-4, SD 0.41). Their overall confidence in NLP knowledge significantly increased (P=.002) from 1.35 (SD 0.47) to 2.79 (SD 0.46). Confidence in coding abilities also improved significantly (P=.01), increasing from 1.33 (SD 0.60) to 2.25 (SD 0.42). Open-ended feedback suggested incorporating thematic analysis and exploring additional datasets for future workshops.</p><p><strong>Conclusions: </strong>This study illustrates the effectiveness of a tailored data science workshop for trainees in psychiatry and psychology, focusing on applying NLP techniques for suicide prevention research. The workshop significantly enhanced
背景:使用人工智能(AI)来分析医疗保健数据在行为健康科学中已经变得很普遍。然而,精神卫生专业人员缺乏培训机会限制了临床医生在临床环境中采用人工智能的能力。人工智能教育对受训者至关重要,使他们具备在实践中应用人工智能工具所需的素养,与数据科学家有效合作,并培养具有计算技能的跨学科研究人员的技能。目的:作为宾夕法尼亚大学自杀预防实施创新研究中心的一部分,我们开发、实施并评估了一个虚拟研讨会,以教育精神病学和心理学学员使用人工智能进行自杀预防研究。方法:研讨会在安全的Microsoft Azure Databricks云计算和分析环境中使用Jupyter笔记本向学员介绍自然语言处理(NLP)概念和Python编码技能。我们设计了一个3小时的研讨会,涵盖了4个关键的NLP主题:数据表征、数据标准化、概念提取和统计分析。为了展示现实世界的应用,我们对电子健康记录中的主诉进行了处理,以比较不同种族、民族和年龄人群中与自杀相关的遭遇的流行程度。培训材料是基于标准的NLP技术和特定领域的任务开发的,例如预处理与精神病学相关的首字母缩略词。两位研究人员起草并演示了代码,并结合自杀预防实施创新研究方法核心的反馈对材料进行了完善。为了评估研讨会的有效性,我们使用了Kirkpatrick项目评估模型,重点关注参与者的反应(第一级)和学习成果(第二级)。使用配对t检验评估研讨会前后知识和技能的信心变化,并包括开放式问题,以收集反馈以供未来改进。结果:有住院医师、博士后、精神科、心理科研究生共10名学员虚拟参与。参与者发现研讨会有帮助(在1-4的范围内平均3.17,标准差0.41)。他们对NLP知识的整体信心从1.35 (SD 0.47)显著增加到2.79 (SD 0.46) (P= 0.002)。对编码能力的信心也显著提高(P= 0.01),从1.33 (SD 0.60)增加到2.25 (SD 0.42)。开放式反馈建议纳入专题分析并为今后的讲习班探索更多的数据集。结论:本研究说明了为精神病学和心理学学员量身定制的数据科学研讨会的有效性,重点是将NLP技术应用于自杀预防研究。工作坊大大提升参加者进行数据科学研究的信心。未来的研讨会将涵盖更多感兴趣的主题,如使用大型语言模型、专题分析、不同的数据集和多方面的结果。这包括检查参与者的学习如何影响他们的实践和研究,以及通过案例研究等方法评估自我报告信心之外的知识和技能,以获得更深入的见解。
{"title":"Data Science Education for Residents, Researchers, and Students in Psychiatry and Psychology: Program Development and Evaluation Study.","authors":"Hayoung K Donnelly, David Mandell, Sy Hwang, Emily Schriver, Ugurcan Vurgun, Graydon Neill, Esha Patel, Megan E Reilly, Michael Steinberg, Amber Calloway, Robert Gallop, Maria A Oquendo, Gregory K Brown, Danielle L Mowery","doi":"10.2196/75125","DOIUrl":"https://doi.org/10.2196/75125","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;The use of artificial intelligence (AI) to analyze health care data has become common in behavioral health sciences. However, the lack of training opportunities for mental health professionals limits clinicians' ability to adopt AI in clinical settings. AI education is essential for trainees, equipping them with the literacy needed to implement AI tools in practice, collaborate effectively with data scientists, and develop skills as interdisciplinary researchers with computing skills.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Objective: &lt;/strong&gt;As part of the Penn Innovation in Suicide Prevention Implementation Research Center, we developed, implemented, and evaluated a virtual workshop to educate psychiatry and psychology trainees on using AI for suicide prevention research.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;The workshop introduced trainees to natural language processing (NLP) concepts and Python coding skills using Jupyter notebooks within a secure Microsoft Azure Databricks cloud computing and analytics environment. We designed a 3-hour workshop that covered 4 key NLP topics: data characterization, data standardization, concept extraction, and statistical analysis. To demonstrate real-world applications, we processed chief complaints from electronic health records to compare the prevalence of suicide-related encounters across populations by race, ethnicity, and age. Training materials were developed based on standard NLP techniques and domain-specific tasks, such as preprocessing psychiatry-related acronyms. Two researchers drafted and demonstrated the code, incorporating feedback from the Methods Core of the Innovation in Suicide Prevention Implementation Research to refine the materials. To evaluate the effectiveness of the workshop, we used the Kirkpatrick program evaluation model, focusing on participants' reactions (level 1) and learning outcomes (level 2). Confidence changes in knowledge and skills before and after the workshop were assessed using paired t tests, and open-ended questions were included to gather feedback for future improvements.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;A total of 10 trainees participated in the workshop virtually, including residents, postdoctoral researchers, and graduate students from the psychiatry and psychology departments. The participants found the workshop helpful (mean 3.17 on a scale of 1-4, SD 0.41). Their overall confidence in NLP knowledge significantly increased (P=.002) from 1.35 (SD 0.47) to 2.79 (SD 0.46). Confidence in coding abilities also improved significantly (P=.01), increasing from 1.33 (SD 0.60) to 2.25 (SD 0.42). Open-ended feedback suggested incorporating thematic analysis and exploring additional datasets for future workshops.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;This study illustrates the effectiveness of a tailored data science workshop for trainees in psychiatry and psychology, focusing on applying NLP techniques for suicide prevention research. The workshop significantly enhanced ","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e75125"},"PeriodicalIF":3.2,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145991329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Implementing Large Language Models to Support Misconception-Based Collaborative Learning in Health Care Education. 在卫生保健教育中实施大型语言模型以支持基于误解的协作学习。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-16 DOI: 10.2196/81875
Brandon C J Cheah, Shefaly Shorey, Jun Hong Ch'ng, Chee Wah Tan

Unlabelled: This paper proposes a framework for leveraging large language models (LLMs) to generate misconceptions as a tool for collaborative learning in health care education. While misconceptions-particularly those generated by AI-are often viewed as detrimental to learning, we present an alternative perspective: that LLM-generated misconceptions, when addressed through structured peer discussion, can promote conceptual change and critical thinking. The paper outlines use cases across health care disciplines, including both clinical and basic science contexts, and a practical 10-step guidance for educators to implement the framework. It also highlights the need for medium- to long-term research to evaluate the impact of LLM-supported learning on student outcomes. This framework may support health care educators globally in integrating emerging AI technologies into their teaching, regardless of the disciplinary focus.

未标记:本文提出了一个框架,利用大型语言模型(llm)来产生误解,作为卫生保健教育协作学习的工具。虽然误解——尤其是人工智能产生的误解——通常被认为对学习有害,但我们提出了另一种观点:法学硕士产生的误解,通过结构化的同行讨论来解决,可以促进观念的改变和批判性思维。该文件概述了跨卫生保健学科的用例,包括临床和基础科学背景,并为教育工作者实施该框架提供了实用的10步指导。它还强调需要进行中长期研究,以评估法学硕士支持的学习对学生成绩的影响。该框架可支持全球卫生保健教育工作者将新兴的人工智能技术纳入其教学,无论其学科重点如何。
{"title":"Implementing Large Language Models to Support Misconception-Based Collaborative Learning in Health Care Education.","authors":"Brandon C J Cheah, Shefaly Shorey, Jun Hong Ch'ng, Chee Wah Tan","doi":"10.2196/81875","DOIUrl":"https://doi.org/10.2196/81875","url":null,"abstract":"<p><strong>Unlabelled: </strong>This paper proposes a framework for leveraging large language models (LLMs) to generate misconceptions as a tool for collaborative learning in health care education. While misconceptions-particularly those generated by AI-are often viewed as detrimental to learning, we present an alternative perspective: that LLM-generated misconceptions, when addressed through structured peer discussion, can promote conceptual change and critical thinking. The paper outlines use cases across health care disciplines, including both clinical and basic science contexts, and a practical 10-step guidance for educators to implement the framework. It also highlights the need for medium- to long-term research to evaluate the impact of LLM-supported learning on student outcomes. This framework may support health care educators globally in integrating emerging AI technologies into their teaching, regardless of the disciplinary focus.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e81875"},"PeriodicalIF":3.2,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145991349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI-Driven Objective Structured Clinical Examination Generation in Digital Health Education: Comparative Analysis of Three GPT-4o Configurations. 数字化健康教育中人工智能驱动的目标结构化临床考试生成:三种gpt - 40配置的比较分析。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-15 DOI: 10.2196/82116
Zineb Zouakia, Emmanuel Logak, Alan Szymczak, Jean-Philippe Jais, Anita Burgun, Rosy Tsopra

Background: Objective Structured Clinical Examinations (OSCEs) are used as an evaluation method in medical education, but require significant pedagogical expertise and investment, especially in emerging fields like digital health. Large language models (LLMs), such as ChatGPT (OpenAI), have shown potential in automating educational content generation. However, OSCE generation using LLMs remains underexplored.

Objective: This study aims to evaluate 3 GPT-4o configurations for generating OSCE stations in digital health: (1) standard GPT with a simple prompt and OSCE guidelines; (2) personalized GPT with a simple prompt, OSCE guidelines, and a reference book in digital health; and (3) simulated-agents GPT with a structured prompt simulating specialized OSCE agents and the digital health reference book.

Methods: Overall, 24 OSCE stations were generated across 8 digital health topics with each GPT-4o configuration. Format compliance was evaluated by one expert, while educational content was assessed independently by 2 digital health experts, blind to GPT-4o configurations, using a comprehensive assessment grid. Statistical analyses were performed using Kruskal-Wallis tests.

Results: Simulated-agents GPT performed best in format compliance and most content quality criteria, including accuracy (mean 4.47/5, SD 0.28; P=.01) and clarity (mean 4.46/5, SD 0.52; P=.004). It also had 88% (14/16) for usability without major revisions and first-place preference ranking, outperforming the other configurations. Personalized GPT showed the lowest format compliance, while standard GPT scored lowest for clarity and educational value.

Conclusions: Structured prompting strategies, particularly agents' simulation, enhance the reliability and usability of LLM-generated OSCE content. These results support the use of artificial intelligence in medical education, while confirming the need for expert validation.

背景:目的结构化临床考试(OSCEs)在医学教育中被用作一种评估方法,但需要大量的教学专业知识和投资,特别是在数字健康等新兴领域。大型语言模型(llm),如ChatGPT (OpenAI),已经显示出自动化教育内容生成的潜力。然而,使用llm生成欧安组织仍未得到充分探索。目的:本研究旨在评估在数字卫生中生成OSCE站点的3种GPT- 40配置:(1)具有简单提示和OSCE指南的标准GPT;(2)个性化GPT,包含简单提示、欧安组织指南和数字健康参考书;(3)模拟代理GPT,具有结构化提示,模拟专门的欧安组织代理和数字健康参考书。方法:总体而言,每个gpt - 40配置在8个数字健康主题中生成24个OSCE站点。格式依从性由一名专家评估,而教育内容由2名数字健康专家独立评估,不受gpt - 40配置的影响,使用综合评估网格。采用Kruskal-Wallis检验进行统计分析。结果:模拟代理GPT在格式遵从性和大多数内容质量标准方面表现最佳,包括准确性(平均4.47/5,SD 0.28; P= 0.01)和清晰度(平均4.46/5,SD 0.52; P= 0.004)。它也有88%(14/16)的可用性,没有重大的修改和首选排名,优于其他配置。个性化GPT表现出最低的格式遵从性,而标准GPT在清晰度和教育价值方面得分最低。结论:结构化提示策略,特别是智能体模拟,提高了llm生成的OSCE内容的可靠性和可用性。这些结果支持在医学教育中使用人工智能,同时确认需要专家验证。
{"title":"AI-Driven Objective Structured Clinical Examination Generation in Digital Health Education: Comparative Analysis of Three GPT-4o Configurations.","authors":"Zineb Zouakia, Emmanuel Logak, Alan Szymczak, Jean-Philippe Jais, Anita Burgun, Rosy Tsopra","doi":"10.2196/82116","DOIUrl":"https://doi.org/10.2196/82116","url":null,"abstract":"<p><strong>Background: </strong>Objective Structured Clinical Examinations (OSCEs) are used as an evaluation method in medical education, but require significant pedagogical expertise and investment, especially in emerging fields like digital health. Large language models (LLMs), such as ChatGPT (OpenAI), have shown potential in automating educational content generation. However, OSCE generation using LLMs remains underexplored.</p><p><strong>Objective: </strong>This study aims to evaluate 3 GPT-4o configurations for generating OSCE stations in digital health: (1) standard GPT with a simple prompt and OSCE guidelines; (2) personalized GPT with a simple prompt, OSCE guidelines, and a reference book in digital health; and (3) simulated-agents GPT with a structured prompt simulating specialized OSCE agents and the digital health reference book.</p><p><strong>Methods: </strong>Overall, 24 OSCE stations were generated across 8 digital health topics with each GPT-4o configuration. Format compliance was evaluated by one expert, while educational content was assessed independently by 2 digital health experts, blind to GPT-4o configurations, using a comprehensive assessment grid. Statistical analyses were performed using Kruskal-Wallis tests.</p><p><strong>Results: </strong>Simulated-agents GPT performed best in format compliance and most content quality criteria, including accuracy (mean 4.47/5, SD 0.28; P=.01) and clarity (mean 4.46/5, SD 0.52; P=.004). It also had 88% (14/16) for usability without major revisions and first-place preference ranking, outperforming the other configurations. Personalized GPT showed the lowest format compliance, while standard GPT scored lowest for clarity and educational value.</p><p><strong>Conclusions: </strong>Structured prompting strategies, particularly agents' simulation, enhance the reliability and usability of LLM-generated OSCE content. These results support the use of artificial intelligence in medical education, while confirming the need for expert validation.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e82116"},"PeriodicalIF":3.2,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145991377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Web-Based Virtual Environment Versus Face-To-Face Delivery for Team-Based Learning of Anesthesia Techniques Among Undergraduate Medical Students: Randomized Controlled Trial. 基于网络的虚拟环境与面对面的方式在本科医学生中进行团队麻醉技术学习:随机对照试验。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-15 DOI: 10.2196/80097
Darunee Sripadungkul, Suhattaya Boonmak, Monsicha Somjit, Narin Plailaharn, Wimonrat Sriraj, Polpun Boonmak
<p><strong>Background: </strong>Foundational knowledge of anesthesia techniques is essential for medical students. Team-based learning (TBL) improves engagement. Web-based virtual environments (WBVEs) allow many learners to join the same session in real time while being guided by an instructor.</p><p><strong>Objective: </strong>This study aimed to compare a WBVE with face-to-face (F2F) delivery of the same TBL curriculum in terms of postclass knowledge and learner satisfaction.</p><p><strong>Methods: </strong>We conducted a randomized, controlled, assessor-blinded trial at a Thai medical school from August 2024 to January 2025. Eligible participants were fifth-year medical students from the Faculty of Medicine, Khon Kaen University, who attended the anesthesiology course at the department of anesthesiology. Students who had previously completed the anesthesiology course or were unable to comply with the study protocol were excluded. They were allocated to one of the groups using a computer-generated sequence, with concealment of allocation to WBVE (on the Spatial platform) or F2F sessions. Both groups received identical 10-section content in a standardized TBL sequence lasting 130 minutes. Only the delivery mode differed (Spatial WBVE vs classroom F2F). The primary outcome was the postclass multiple-choice questionnaire score. The secondary outcome was learner satisfaction. Individual knowledge was assessed before and after the session using a 15-item questionnaire containing multiple-choice questions via Google Forms. Satisfaction was measured immediately after class on a 5-point Likert scale. Outcome scoring and data analysis were blinded to group assignment. Participants and instructors were not blinded.</p><p><strong>Results: </strong>In total, 79 students were randomized in this study (F2F: n=38, 48%; WBVE: n=41, 52%). We excluded 2% (1/41) of the students in the WBVE group due to incomplete data. There were complete data for the analysis for 78 participants (F2F: n=38, 49%; WBVE: n=40, 51%). Preclass scores were similar between groups (F2F: mean 6.03, SD 2.05; WBVE: mean 6.20, SD 2.04). Postclass knowledge did not differ significantly (F2F: mean 11.24, SD 1.93; WBVE: mean 10.40, SD 2.62; mean difference 0.88, 95% CI -0.18 to 1.94; P=.12). Learner satisfaction favored F2F learning across multiple domains, including overall course satisfaction. Overall satisfaction favored F2F learning (mean difference 0.42, 95% CI 0.07-0.77; P=.01). Both groups ran as planned. No adverse events were reported. No technical failures occurred in the WBVE group.</p><p><strong>Conclusions: </strong>In this trial, WBVE-delivered TBL produced similar short-term knowledge gains to F2F delivery, but learner satisfaction was lower in the WBVE group. Unlike many previous studies, this trial compared WBVE and F2F delivery while keeping the TBL curriculum and prespecified outcomes identical across groups. These findings support WBVEs as a scalable option when physical sp
背景:麻醉技术的基础知识是医学生必不可少的。基于团队的学习(TBL)提高了参与度。基于web的虚拟环境(WBVEs)允许许多学习者在讲师的指导下实时加入同一个课程。目的:本研究旨在比较面对面授课与面对面授课在课后知识和学习者满意度方面的差异。方法:我们于2024年8月至2025年1月在泰国一所医学院进行了一项随机、对照、评估者盲法试验。符合条件的参与者是孔敬大学医学院的五年级医学生,他们参加了麻醉科的麻醉学课程。先前已完成麻醉学课程或无法遵守研究方案的学生被排除在外。他们使用计算机生成的序列被分配到其中一个组,隐藏分配到WBVE(在空间平台上)或F2F会议。两组在标准化TBL序列中接受相同的10段内容,持续130分钟。只有交付模式不同(空间WBVE vs教室F2F)。主要结果为课后多项选择问卷得分。次要结果是学习者满意度。在课程前后,通过谷歌表格使用包含多项选择题的15项问卷来评估个人知识。满意度在下课后立即以5分李克特量表进行测量。结果评分和数据分析采用分组盲法。参与者和指导员没有被蒙蔽。结果:本研究共随机纳入79名学生(F2F: n=38, 48%; WBVE: n=41, 52%)。由于数据不完整,我们将2%(1/41)的WBVE组学生排除在外。78名受试者(F2F: n= 38,49%; WBVE: n= 40,51%)有完整的分析数据。两组间课前评分相近(F2F: mean 6.03, SD 2.05; WBVE: mean 6.20, SD 2.04)。课后知识差异无统计学意义(F2F: mean 11.24, SD 1.93; WBVE: mean 10.40, SD 2.62;平均差异0.88,95% CI -0.18 ~ 1.94; P= 0.12)。学习者满意度有利于跨多个领域的F2F学习,包括总体课程满意度。总体满意度倾向于F2F学习(平均差异0.42,95% CI 0.07-0.77; P= 0.01)。两组都按照计划进行。无不良事件报告。WBVE组未发生技术故障。结论:在本试验中,WBVE提供的TBL与F2F提供的TBL产生了相似的短期知识收益,但WBVE组的学习者满意度较低。与之前的许多研究不同,该试验比较了WBVE和F2F的交付,同时保持了TBL课程和预先指定的结果在各组之间相同。这些发现表明,当存在物理空间、学习者数量或限制时,wbve是一种可扩展的选择。然而,在WBVE中较低的满意度突出表明,在广泛实施之前,现实世界需要改进便利、用户体验设计和技术准备。试验注册:泰国临床试验注册中心TCTR20240708012;https://www.thaiclinicaltrials.org/show/TCTR20240708012。
{"title":"Web-Based Virtual Environment Versus Face-To-Face Delivery for Team-Based Learning of Anesthesia Techniques Among Undergraduate Medical Students: Randomized Controlled Trial.","authors":"Darunee Sripadungkul, Suhattaya Boonmak, Monsicha Somjit, Narin Plailaharn, Wimonrat Sriraj, Polpun Boonmak","doi":"10.2196/80097","DOIUrl":"https://doi.org/10.2196/80097","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;Foundational knowledge of anesthesia techniques is essential for medical students. Team-based learning (TBL) improves engagement. Web-based virtual environments (WBVEs) allow many learners to join the same session in real time while being guided by an instructor.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Objective: &lt;/strong&gt;This study aimed to compare a WBVE with face-to-face (F2F) delivery of the same TBL curriculum in terms of postclass knowledge and learner satisfaction.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;We conducted a randomized, controlled, assessor-blinded trial at a Thai medical school from August 2024 to January 2025. Eligible participants were fifth-year medical students from the Faculty of Medicine, Khon Kaen University, who attended the anesthesiology course at the department of anesthesiology. Students who had previously completed the anesthesiology course or were unable to comply with the study protocol were excluded. They were allocated to one of the groups using a computer-generated sequence, with concealment of allocation to WBVE (on the Spatial platform) or F2F sessions. Both groups received identical 10-section content in a standardized TBL sequence lasting 130 minutes. Only the delivery mode differed (Spatial WBVE vs classroom F2F). The primary outcome was the postclass multiple-choice questionnaire score. The secondary outcome was learner satisfaction. Individual knowledge was assessed before and after the session using a 15-item questionnaire containing multiple-choice questions via Google Forms. Satisfaction was measured immediately after class on a 5-point Likert scale. Outcome scoring and data analysis were blinded to group assignment. Participants and instructors were not blinded.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;In total, 79 students were randomized in this study (F2F: n=38, 48%; WBVE: n=41, 52%). We excluded 2% (1/41) of the students in the WBVE group due to incomplete data. There were complete data for the analysis for 78 participants (F2F: n=38, 49%; WBVE: n=40, 51%). Preclass scores were similar between groups (F2F: mean 6.03, SD 2.05; WBVE: mean 6.20, SD 2.04). Postclass knowledge did not differ significantly (F2F: mean 11.24, SD 1.93; WBVE: mean 10.40, SD 2.62; mean difference 0.88, 95% CI -0.18 to 1.94; P=.12). Learner satisfaction favored F2F learning across multiple domains, including overall course satisfaction. Overall satisfaction favored F2F learning (mean difference 0.42, 95% CI 0.07-0.77; P=.01). Both groups ran as planned. No adverse events were reported. No technical failures occurred in the WBVE group.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;In this trial, WBVE-delivered TBL produced similar short-term knowledge gains to F2F delivery, but learner satisfaction was lower in the WBVE group. Unlike many previous studies, this trial compared WBVE and F2F delivery while keeping the TBL curriculum and prespecified outcomes identical across groups. These findings support WBVEs as a scalable option when physical sp","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e80097"},"PeriodicalIF":3.2,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145991397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptation of the Japanese Version of the 12-Item Attitudes Towards Artificial Intelligence Scale for Medical Trainees: Multicenter Development and Validation Study. 医学实习生对人工智能12项态度量表日文版的改编:多中心开发与验证研究
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-14 DOI: 10.2196/81986
Hirohisa Fujikawa, Hirotake Mori, Kayo Kondo, Yuji Nishizaki, Yuichiro Yano, Toshio Naito
<p><strong>Background: </strong>In the current era of artificial intelligence (AI), use of AI has increased in both clinical practice and medical education. Nevertheless, it is probable that perspectives on the prospects and risks of AI vary among individuals. Given the potential for attitudes toward AI to significantly influence its integration into medical practice and educational initiatives, it is essential to assess these attitudes using a validated tool. The recently developed 12-item Attitudes Towards Artificial Intelligence scale has demonstrated good validity and reliability for the general population, suggesting its potential for extensive use in future studies. However, to our knowledge, there is currently no validated Japanese version of the scale. The lack of a Japanese version hinders research and educational efforts aimed at understanding and improving AI integration into the Japanese health care and medical education system.</p><p><strong>Objective: </strong>We aimed to develop the Japanese version of the 12-item Attitudes Towards Artificial Intelligence scale (J-ATTARI-12) and investigate whether it is applicable to medical trainees.</p><p><strong>Methods: </strong>We first translated the original English-language scale into Japanese. To examine its psychometric properties, we then conducted a validation survey by distributing the translated version as an online questionnaire to medical students and residents across Japan from June 2025 to July 2025. We assessed structural validity through factor analysis and convergent validity by computing the Pearson correlation coefficient between the J-ATTARI-12 scores and scores on attitude toward robots. Internal consistency reliability was assessed using Cronbach α values.</p><p><strong>Results: </strong>We included 326 participants in our analysis. We used a split-half validation approach, with exploratory factor analysis (EFA) on the first half and confirmatory factor analysis on the second half. EFA suggested a 2-factor solution (factor 1: AI anxiety and aversion; factor 2: AI optimism and acceptance). Confirmatory factor analysis revealed that the model fitness indexes of the 2-factor structure suggested by the EFA were good (comparative fit index=0.914 [>0.900]; root mean square error of approximation=0.075 [<0.080]; standardized root mean square residual=0.056 [<0.080]) and superior to those of the 1-factor structure. The value of the Pearson correlation coefficient between the J-ATTARI-12 scores and the attitude toward robots scores was 0.52, which indicated good convergent validity. The Cronbach α for all 12 items was 0.84, which indicated a high level of internal consistency reliability.</p><p><strong>Conclusions: </strong>We developed and validated the J-ATTARI-12. The developed instrument had good structural validity, convergent validity, and internal consistency reliability for medical trainees. The J-ATTARI-12 is expected to stimulate future studies and educational initiative
背景:在当前人工智能(AI)时代,人工智能在临床实践和医学教育中的应用都有所增加。然而,对人工智能的前景和风险的看法可能因人而异。鉴于人们对人工智能的态度可能会对其融入医疗实践和教育举措产生重大影响,有必要使用一种经过验证的工具来评估这些态度。最近开发的12项人工智能态度量表在一般人群中显示出良好的效度和信度,表明其在未来研究中有广泛应用的潜力。然而,据我们所知,目前还没有经过验证的日文量表。日本版本的缺乏阻碍了旨在理解和改善人工智能融入日本医疗保健和医学教育体系的研究和教育努力。目的:编制日语版12项人工智能态度量表(J-ATTARI-12),探讨其是否适用于医学实习生。方法:首先将原英语量表翻译成日语。为了检验其心理测量特性,我们在2025年6月至2025年7月期间将翻译版本作为在线问卷分发给日本各地的医科学生和居民,进行了验证性调查。我们通过因子分析评估结构效度,并通过计算J-ATTARI-12得分与机器人态度得分之间的Pearson相关系数来评估收敛效度。采用Cronbach α值评估内部一致性信度。结果:我们纳入了326名参与者。我们使用了二分验证方法,探索性因子分析(EFA)在前半部分,验证性因子分析在后半部分。EFA提出了一个双因素解决方案(因素1:人工智能焦虑和厌恶;因素2:人工智能乐观和接受)。验证性因子分析显示,EFA建议的2因素结构模型适应度指标较好(比较拟合指数=0.914[>0.900],近似均方根误差=0.075]。该量表具有良好的结构效度、收敛效度和内部一致性信度。J-ATTARI-12预计将刺激未来的研究和教育举措,有效评估和加强人工智能与临床实践和医学教育系统的整合。
{"title":"Adaptation of the Japanese Version of the 12-Item Attitudes Towards Artificial Intelligence Scale for Medical Trainees: Multicenter Development and Validation Study.","authors":"Hirohisa Fujikawa, Hirotake Mori, Kayo Kondo, Yuji Nishizaki, Yuichiro Yano, Toshio Naito","doi":"10.2196/81986","DOIUrl":"https://doi.org/10.2196/81986","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;In the current era of artificial intelligence (AI), use of AI has increased in both clinical practice and medical education. Nevertheless, it is probable that perspectives on the prospects and risks of AI vary among individuals. Given the potential for attitudes toward AI to significantly influence its integration into medical practice and educational initiatives, it is essential to assess these attitudes using a validated tool. The recently developed 12-item Attitudes Towards Artificial Intelligence scale has demonstrated good validity and reliability for the general population, suggesting its potential for extensive use in future studies. However, to our knowledge, there is currently no validated Japanese version of the scale. The lack of a Japanese version hinders research and educational efforts aimed at understanding and improving AI integration into the Japanese health care and medical education system.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Objective: &lt;/strong&gt;We aimed to develop the Japanese version of the 12-item Attitudes Towards Artificial Intelligence scale (J-ATTARI-12) and investigate whether it is applicable to medical trainees.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;We first translated the original English-language scale into Japanese. To examine its psychometric properties, we then conducted a validation survey by distributing the translated version as an online questionnaire to medical students and residents across Japan from June 2025 to July 2025. We assessed structural validity through factor analysis and convergent validity by computing the Pearson correlation coefficient between the J-ATTARI-12 scores and scores on attitude toward robots. Internal consistency reliability was assessed using Cronbach α values.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;We included 326 participants in our analysis. We used a split-half validation approach, with exploratory factor analysis (EFA) on the first half and confirmatory factor analysis on the second half. EFA suggested a 2-factor solution (factor 1: AI anxiety and aversion; factor 2: AI optimism and acceptance). Confirmatory factor analysis revealed that the model fitness indexes of the 2-factor structure suggested by the EFA were good (comparative fit index=0.914 [&gt;0.900]; root mean square error of approximation=0.075 [&lt;0.080]; standardized root mean square residual=0.056 [&lt;0.080]) and superior to those of the 1-factor structure. The value of the Pearson correlation coefficient between the J-ATTARI-12 scores and the attitude toward robots scores was 0.52, which indicated good convergent validity. The Cronbach α for all 12 items was 0.84, which indicated a high level of internal consistency reliability.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;We developed and validated the J-ATTARI-12. The developed instrument had good structural validity, convergent validity, and internal consistency reliability for medical trainees. The J-ATTARI-12 is expected to stimulate future studies and educational initiative","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e81986"},"PeriodicalIF":3.2,"publicationDate":"2026-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145991274","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation of a Problem-Based Learning Program's Effect on Artificial Intelligence Ethics Among Japanese Medical Students: Mixed Methods Study. 评价基于问题的学习计划对日本医学生人工智能伦理的影响:混合方法研究。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-14 DOI: 10.2196/84535
Yuma Ota, Yoshikazu Asada, Saori Kubo, Takeshi Kanno, Machiko Saeki Yagi, Yasushi Matsuyama
<p><strong>Background: </strong>The rapid advancement of artificial intelligence (AI) has had a substantial impact on medicine, necessitating the integration of AI education into medical school curricula. However, such integration remains limited. A key challenge is the discrepancy between medical students' positive perceptions of AI and their actual competencies, with research in Japan identifying specific gaps in the students' competencies in understanding regulations and discussing ethical issues.</p><p><strong>Objective: </strong>This study evaluates the effectiveness of an educational program designed to improve medical students' competencies in understanding legal and ethical AI-related issues. It addresses the following research questions: (1) Does this educational program improve students' knowledge of AI and its legal and ethical issues, and what is each program element's contribution to this knowledge? (2) How does this educational program qualitatively change medical students' thoughts on these issues from an abstract understanding to a concrete and structured thought process?</p><p><strong>Methods: </strong>This mixed methods study used a single-group pretest and posttest framework involving 118 fourth-year medical students. The 1-day intervention comprised a lecture and problem-based learning (PBL) session centered on a clinical case. A 24-item multiple-choice questionnaire (MCQ) was administered at 3 time points (pretest, midtest, and posttest), and descriptive essays were collected before and after the intervention. Data were analyzed using linear mixed-effects models, the Wilcoxon signed-rank test, and text mining, including comparative frequency analysis and cooccurrence network analysis with Jaccard coefficients. An optional survey on student perceptions based on the attention, relevance, confidence, and satisfaction model was conducted (n=76, 64.4%).</p><p><strong>Results: </strong>Objective knowledge scores increased significantly from the pretest (median 17, IQR 15-18) to posttest (median 19, IQR 17-21; β=1.42; P<.001). No significant difference was observed between score gains during the lecture and PBL phases (P=.54). Qualitative text analysis revealed the significant transformation of cooccurrence network structures (Jaccard coefficients 0.116 and 0.121) from fragmented clusters to integrated networks. Students also used professional and ethical terminology more frequently. For instance, use of the term "bias" in patient explanations increased from 10 (8.5%) at pretest to 25 (21.2%) at posttest, while references to "personal information" in physician precautions increased from 36 (30.5%) to 50 (42.4%). The optional survey indicated that students' confidence (mean 3.78, SD 0.87) was significantly lower than their perception of the program's relevance (mean 4.20, SD 0.71; P<.001).</p><p><strong>Conclusions: </strong>This PBL-based program was associated with the improvements in knowledge and, more importantly, a structural t
背景:人工智能(AI)的快速发展对医学产生了重大影响,有必要将人工智能教育纳入医学院课程。然而,这种整合仍然有限。一项关键挑战是医学生对人工智能的积极看法与他们的实际能力之间存在差异,日本的研究确定了学生在理解法规和讨论道德问题方面的能力存在具体差距。目的:本研究评估一项旨在提高医学生理解人工智能相关法律和伦理问题能力的教育计划的有效性。它解决了以下研究问题:(1)这个教育项目是否提高了学生对人工智能及其法律和伦理问题的认识,每个项目元素对这一知识的贡献是什么?(2)这一教育项目如何从本质上改变医学生对这些问题的看法,从抽象的理解转变为具体的、有组织的思维过程?方法:本研究采用单组前测和后测框架,纳入118名四年级医学生。为期1天的干预包括以临床病例为中心的讲座和基于问题的学习(PBL)会议。在3个时间点(测试前、测试中和测试后)进行24项选择问卷(MCQ),并在干预前后收集描述性文章。数据分析使用线性混合效应模型、Wilcoxon符号秩检验和文本挖掘,包括比较频率分析和Jaccard系数共现网络分析。基于注意、关联、信心和满意度模型对学生的认知进行了选择性调查(n=76, 64.4%)。结果:客观知识得分从测试前(中位数17,IQR 15-18)到测试后(中位数19,IQR 17-21)显著提高(β=1.42)。结论:基于pbl的项目与知识的提高有关,更重要的是,学生对人工智能伦理的思考从抽象层面转变为具体的、有临床依据的推理。定量和定性结果之间的差异表明mcq在评估PBL培养的高阶技能方面存在局限性。总体而言,本研究表明PBL作为一种有效的人工智能伦理教育教学方法的潜力。
{"title":"Evaluation of a Problem-Based Learning Program's Effect on Artificial Intelligence Ethics Among Japanese Medical Students: Mixed Methods Study.","authors":"Yuma Ota, Yoshikazu Asada, Saori Kubo, Takeshi Kanno, Machiko Saeki Yagi, Yasushi Matsuyama","doi":"10.2196/84535","DOIUrl":"https://doi.org/10.2196/84535","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;The rapid advancement of artificial intelligence (AI) has had a substantial impact on medicine, necessitating the integration of AI education into medical school curricula. However, such integration remains limited. A key challenge is the discrepancy between medical students' positive perceptions of AI and their actual competencies, with research in Japan identifying specific gaps in the students' competencies in understanding regulations and discussing ethical issues.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Objective: &lt;/strong&gt;This study evaluates the effectiveness of an educational program designed to improve medical students' competencies in understanding legal and ethical AI-related issues. It addresses the following research questions: (1) Does this educational program improve students' knowledge of AI and its legal and ethical issues, and what is each program element's contribution to this knowledge? (2) How does this educational program qualitatively change medical students' thoughts on these issues from an abstract understanding to a concrete and structured thought process?&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;This mixed methods study used a single-group pretest and posttest framework involving 118 fourth-year medical students. The 1-day intervention comprised a lecture and problem-based learning (PBL) session centered on a clinical case. A 24-item multiple-choice questionnaire (MCQ) was administered at 3 time points (pretest, midtest, and posttest), and descriptive essays were collected before and after the intervention. Data were analyzed using linear mixed-effects models, the Wilcoxon signed-rank test, and text mining, including comparative frequency analysis and cooccurrence network analysis with Jaccard coefficients. An optional survey on student perceptions based on the attention, relevance, confidence, and satisfaction model was conducted (n=76, 64.4%).&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;Objective knowledge scores increased significantly from the pretest (median 17, IQR 15-18) to posttest (median 19, IQR 17-21; β=1.42; P&lt;.001). No significant difference was observed between score gains during the lecture and PBL phases (P=.54). Qualitative text analysis revealed the significant transformation of cooccurrence network structures (Jaccard coefficients 0.116 and 0.121) from fragmented clusters to integrated networks. Students also used professional and ethical terminology more frequently. For instance, use of the term \"bias\" in patient explanations increased from 10 (8.5%) at pretest to 25 (21.2%) at posttest, while references to \"personal information\" in physician precautions increased from 36 (30.5%) to 50 (42.4%). The optional survey indicated that students' confidence (mean 3.78, SD 0.87) was significantly lower than their perception of the program's relevance (mean 4.20, SD 0.71; P&lt;.001).&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;This PBL-based program was associated with the improvements in knowledge and, more importantly, a structural t","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e84535"},"PeriodicalIF":3.2,"publicationDate":"2026-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145985167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Interactive, Image-Based Modules as a Complement to Prosection-Based Anatomy Laboratories: Multicohort Evaluation. 交互式的,基于图像的模块作为对基于检控的解剖实验室的补充:多队列评估。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-13 DOI: 10.2196/85028
Caroline Sumner, Sami L Case, Samuel Franklin, Kristen Platt

Background: As medical and allied health curricula adapt to increasing time constraints, ethical considerations, and resource limitations, digital innovations are becoming vital supplements to donor-based anatomy instruction. While prior studies have examined the effectiveness of prosection versus dissection and the role of digital tools in anatomy learning, few resources align interactive digital modules directly with hands-on prosection experiences.

Objective: This project addresses that gap by introducing an integrated, curriculum-aligned platform for self-guided cadaveric learning.

Methods: We created Anatomy Interactives, a web-based laboratory manual structured to complement prosection laboratories for MD, DPT, and PA students. Modules were developed using iSpring Suite (iSpring Solutions Incorporated) and included interactive labeled images, donor photographs, and quiz-style self-assessments. Learners engaged with modules before, during, or after laboratory sessions. PA/DPT and MD students completed postcourse surveys evaluating module use and perceived impact. MD student examination scores from a 2023 cohort (no module access) were compared to a 2024 cohort (with access) to evaluate effectiveness.

Results: A total of 147 students completed the survey (31 PA/DPT and 116 MD). The majority reported using modules for 1-2 hours per week and found them helpful for both written and laboratory examinations. MD students in the 2024 cohort performed better on all 3 examinations compared to the 2023 cohort, with 2 examination median differences reaching statistical significance (Mann-Whitney U, P<.001). Qualitative feedback highlighted accessibility, content reinforcement, and user engagement as key benefits.

Conclusions: Interactive modules integrated with prosection laboratories enhanced learner engagement and performance. This hybrid digital-donor model shows promise for scalable, learner-centered gross anatomy education.

背景:随着医学和相关卫生课程适应越来越多的时间限制、伦理考虑和资源限制,数字创新正成为以供体为基础的解剖学教学的重要补充。虽然先前的研究已经检查了检控与解剖的有效性以及数字工具在解剖学学习中的作用,但很少有资源将交互式数字模块直接与动手检控经验结合起来。目的:该项目通过引入一个集成的、与课程相一致的自我引导尸体学习平台来解决这一差距。方法:我们创建了Anatomy interactive,这是一个基于网络的实验室手册,用于补充MD, DPT和PA学生的检检实验室。模块使用isspring Suite (isspring Solutions Incorporated)开发,包括交互式标记图像、供体照片和测验式自我评估。学习者在实验之前、期间或之后都参与了模块的学习。PA/DPT和MD学生完成了评估模块使用和感知影响的课后调查。将2023年队列(无模块访问)的MD学生考试分数与2024年队列(有模块访问)的MD学生考试分数进行比较,以评估有效性。结果:共147名学生完成调查,其中PA/DPT 31名,MD 116名。大多数人报告每周使用1-2小时的模块,并发现它们对笔试和实验室考试都很有帮助。与2023队列相比,2024队列的MD学生在所有3项考试中表现更好,其中2项考试中位数差异达到统计学意义(Mann-Whitney U, p)。结论:与检检实验室集成的互动模块提高了学习者的参与度和表现。这种混合数字捐赠模式显示了可扩展的、以学习者为中心的大体解剖学教育的前景。
{"title":"Interactive, Image-Based Modules as a Complement to Prosection-Based Anatomy Laboratories: Multicohort Evaluation.","authors":"Caroline Sumner, Sami L Case, Samuel Franklin, Kristen Platt","doi":"10.2196/85028","DOIUrl":"https://doi.org/10.2196/85028","url":null,"abstract":"<p><strong>Background: </strong>As medical and allied health curricula adapt to increasing time constraints, ethical considerations, and resource limitations, digital innovations are becoming vital supplements to donor-based anatomy instruction. While prior studies have examined the effectiveness of prosection versus dissection and the role of digital tools in anatomy learning, few resources align interactive digital modules directly with hands-on prosection experiences.</p><p><strong>Objective: </strong>This project addresses that gap by introducing an integrated, curriculum-aligned platform for self-guided cadaveric learning.</p><p><strong>Methods: </strong>We created Anatomy Interactives, a web-based laboratory manual structured to complement prosection laboratories for MD, DPT, and PA students. Modules were developed using iSpring Suite (iSpring Solutions Incorporated) and included interactive labeled images, donor photographs, and quiz-style self-assessments. Learners engaged with modules before, during, or after laboratory sessions. PA/DPT and MD students completed postcourse surveys evaluating module use and perceived impact. MD student examination scores from a 2023 cohort (no module access) were compared to a 2024 cohort (with access) to evaluate effectiveness.</p><p><strong>Results: </strong>A total of 147 students completed the survey (31 PA/DPT and 116 MD). The majority reported using modules for 1-2 hours per week and found them helpful for both written and laboratory examinations. MD students in the 2024 cohort performed better on all 3 examinations compared to the 2023 cohort, with 2 examination median differences reaching statistical significance (Mann-Whitney U, P<.001). Qualitative feedback highlighted accessibility, content reinforcement, and user engagement as key benefits.</p><p><strong>Conclusions: </strong>Interactive modules integrated with prosection laboratories enhanced learner engagement and performance. This hybrid digital-donor model shows promise for scalable, learner-centered gross anatomy education.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e85028"},"PeriodicalIF":3.2,"publicationDate":"2026-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145967279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GPT-4o and OpenAI o1 Performance on the 2024 Spanish Competitive Medical Specialty Access Examination: Cross-Sectional Quantitative Evaluation Study. gpt - 40和OpenAI 01在2024年西班牙竞争性医学专业准入考试中的表现:横断面定量评估研究。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-12 DOI: 10.2196/75452
Pau Benito, Mikel Isla-Jover, Pablo González-Castro, Pedro José Fernández Esparcia, Manuel Carpio, Iván Blay-Simón, Pablo Gutiérrez-Bedia, Maria J Lapastora, Beatriz Carratalá, Carlos Carazo-Casas
<p><strong>Background: </strong>In recent years, generative artificial intelligence and large language models (LLMs) have rapidly advanced, offering significant potential to transform medical education. Several studies have evaluated the performance of chatbots on multiple-choice medical examinations.</p><p><strong>Objective: </strong>The study aims to assess the performance of two LLMs-GPT-4o and OpenAI o1-on the Médico Interno Residente (MIR) 2024 examination, the Spanish national medical test that determines eligibility for competitive medical specialist training positions.</p><p><strong>Methods: </strong>A total of 176 questions from the MIR 2024 examination were analyzed. Each question was presented individually to the chatbots to ensure independence and prevent memory retention bias. No additional prompts were introduced to minimize potential bias. For each LLM, response consistency under verification prompting was assessed by systematically asking, "Are you sure?" after each response. Accuracy was defined as the percentage of correct responses compared to the official answers provided by the Spanish Ministry of Health. It was assessed for GPT-4o, OpenAI o1, and, as a benchmark, for a consensus of medical specialists and for the average MIR candidate. Subanalyses included performance across different medical subjects, question difficulty (quintiles based on the percentage of examinees correctly answering each question), and question types (clinical cases vs theoretical questions; positive vs negative questions).</p><p><strong>Results: </strong>Overall accuracy was 89.8% (158/176) for GPT-4o and 90% (160/176) after verification prompting, 92.6% (163/176) for OpenAI o1 and 93.2% (164/176) after verification prompting, 94.3% (166/176) for the consensus of medical specialists, and 56.6% (100/176) for the average MIR candidate. Both LLMs and the consensus of medical specialists outperformed the average MIR candidate across all 20 medical subjects analyzed, with ≥80% LLMs' accuracy in most domains. A performance gradient was observed: LLMs' accuracy gradually declined as question difficulty increased. Slightly higher accuracy was observed for clinical cases compared to theoretical questions, as well as for positive questions compared to negative ones. Both models demonstrated high response consistency, with near-perfect agreement between initial responses and those after the verification prompting.</p><p><strong>Conclusions: </strong>These findings highlight the excellent performance of GPT-4o and OpenAI o1 on the MIR 2024 examination, demonstrating consistent accuracy across medical subjects and question types. The integration of LLMs into medical education presents promising opportunities and is likely to reshape how students prepare for licensing examinations and change our understanding of medical education. Further research should explore how the wording, language, prompting techniques, and image-based questions can influence LLMs' accuracy,
背景:近年来,生成式人工智能和大型语言模型(llm)迅速发展,为改变医学教育提供了巨大的潜力。有几项研究评估了聊天机器人在多项选择医学考试中的表现。目的:这项研究的目的是评估的性能两个LLMs-GPT-4o和OpenAI o1-on的医生Interno Residente (MIR) 2024考试,西班牙国家医学考试决定竞争力的医学专家培训资格的位置。方法:对MIR 2024考题176道题进行分析。每个问题都单独呈现给聊天机器人,以确保独立性,防止记忆偏差。没有引入额外的提示以尽量减少潜在的偏差。对于每个LLM,在验证提示下的响应一致性通过在每个响应后系统地询问“Are you sure?”来评估。准确性定义为与西班牙卫生部提供的官方答案相比,正确回答的百分比。对gpt - 40、OpenAI 01进行了评估,并作为基准,对医学专家的共识和平均MIR候选人进行了评估。亚分析包括不同医学科目的表现、问题难度(基于考生正确回答每个问题的百分比的五分之一)和问题类型(临床案例与理论问题;积极问题与消极问题)。结果:gpt - 40的总体准确率为89.8%(158/176),验证提示后的准确率为90% (160/176),OpenAI 0的总体准确率为92.6%(163/176),验证提示后的准确率为93.2%(164/176),医学专家共识的准确率为94.3% (166/176),MIR候选人的平均准确率为56.6%(100/176)。在所有分析的20个医学科目中,法学硕士和医学专家的共识都优于MIR候选人的平均表现,法学硕士在大多数领域的准确率≥80%。我们观察到一个性能梯度:随着问题难度的增加,LLMs的准确率逐渐下降。与理论问题相比,临床病例的准确性略高,与消极问题相比,积极问题的准确性略高。两种模型均表现出较高的响应一致性,初始响应与验证提示后的响应几乎完全一致。结论:这些发现突出了gpt - 40和OpenAI 01在MIR 2024考试中的优异表现,在医学科目和问题类型中表现出一致的准确性。法学硕士与医学教育的整合提供了有希望的机会,可能会重塑学生准备执照考试的方式,并改变我们对医学教育的理解。进一步的研究应该探索措辞、语言、提示技术和基于图像的问题如何影响法学硕士的准确性,以及评估新兴人工智能模型在类似评估中的表现。
{"title":"GPT-4o and OpenAI o1 Performance on the 2024 Spanish Competitive Medical Specialty Access Examination: Cross-Sectional Quantitative Evaluation Study.","authors":"Pau Benito, Mikel Isla-Jover, Pablo González-Castro, Pedro José Fernández Esparcia, Manuel Carpio, Iván Blay-Simón, Pablo Gutiérrez-Bedia, Maria J Lapastora, Beatriz Carratalá, Carlos Carazo-Casas","doi":"10.2196/75452","DOIUrl":"10.2196/75452","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;In recent years, generative artificial intelligence and large language models (LLMs) have rapidly advanced, offering significant potential to transform medical education. Several studies have evaluated the performance of chatbots on multiple-choice medical examinations.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Objective: &lt;/strong&gt;The study aims to assess the performance of two LLMs-GPT-4o and OpenAI o1-on the Médico Interno Residente (MIR) 2024 examination, the Spanish national medical test that determines eligibility for competitive medical specialist training positions.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;A total of 176 questions from the MIR 2024 examination were analyzed. Each question was presented individually to the chatbots to ensure independence and prevent memory retention bias. No additional prompts were introduced to minimize potential bias. For each LLM, response consistency under verification prompting was assessed by systematically asking, \"Are you sure?\" after each response. Accuracy was defined as the percentage of correct responses compared to the official answers provided by the Spanish Ministry of Health. It was assessed for GPT-4o, OpenAI o1, and, as a benchmark, for a consensus of medical specialists and for the average MIR candidate. Subanalyses included performance across different medical subjects, question difficulty (quintiles based on the percentage of examinees correctly answering each question), and question types (clinical cases vs theoretical questions; positive vs negative questions).&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;Overall accuracy was 89.8% (158/176) for GPT-4o and 90% (160/176) after verification prompting, 92.6% (163/176) for OpenAI o1 and 93.2% (164/176) after verification prompting, 94.3% (166/176) for the consensus of medical specialists, and 56.6% (100/176) for the average MIR candidate. Both LLMs and the consensus of medical specialists outperformed the average MIR candidate across all 20 medical subjects analyzed, with ≥80% LLMs' accuracy in most domains. A performance gradient was observed: LLMs' accuracy gradually declined as question difficulty increased. Slightly higher accuracy was observed for clinical cases compared to theoretical questions, as well as for positive questions compared to negative ones. Both models demonstrated high response consistency, with near-perfect agreement between initial responses and those after the verification prompting.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;These findings highlight the excellent performance of GPT-4o and OpenAI o1 on the MIR 2024 examination, demonstrating consistent accuracy across medical subjects and question types. The integration of LLMs into medical education presents promising opportunities and is likely to reshape how students prepare for licensing examinations and change our understanding of medical education. Further research should explore how the wording, language, prompting techniques, and image-based questions can influence LLMs' accuracy,","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e75452"},"PeriodicalIF":3.2,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12795474/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145960564","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ultrasound-Guided Regional Anesthesia in a Resource-Limited Hospital: Prospective Pilot Study of a Hybrid Training Program. 超声引导区域麻醉在危地马拉资源有限的设置:混合训练模式的前瞻性评估。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2026-01-08 DOI: 10.2196/84181
Jakob E Gamboa, Inge Tamm-Daniels, Roland Flores, Nancy G Sarat Diaz, Mario A Villasenor, Mitchell A Gist, Aidan B Hoie, Christopher Kurinec, Colby G Simmons

Background: Ultrasound-guided regional anesthesia (UGRA) remains underused in low- and middle-income countries due to barriers to training and equipment. Recent advances in portable ultrasound devices and international partnerships have expanded access to UGRA, enhancing patient safety and quality of care.

Objective: This study describes the development and outcomes of a hybrid UGRA training program for anesthesiologists at the Hospital Nacional de Coatepeque (HNC) in Guatemala.

Methods: An educational pilot program for UGRA was developed based on local needs and feedback, comprising 4 weeks of online modules, an in-person educational conference, and 1 month of supervised clinical practice. Evaluation followed the Kirkpatrick framework using preprogram and postprogram surveys adapted from the Global Regional Anesthesia Curricular Engagement model. Outcomes included participants' satisfaction, change in knowledge and skill, and procedural performance. Knowledge and skill assessments were compared before and after the training, and clinical data were recorded for 10 months. Nonparametric tests were used to assess changes and associations with performance outcomes.

Results: All 7 anesthesiologists at HNC completed the training program. Knowledge test scores improved by a median percentage increase of 20.8% (IQR 13.5%-28.1%; r=0.899; P=.02), and procedural skill rating scores increased by a median percentage of 147.1% (IQR 96.9%-197.3%; r=0.904; P=.03) at 1 month and 131.4% (IQR 90.5%-172.3%; r=0.909; P=.04) at 4 months after the program. Participants self-reported high satisfaction and substantial clinical improvement and motivation. A total of 54 peripheral nerve blocks were performed under direct supervision in the first month, with 187 blocks recorded over 10 months. The supraclavicular brachial plexus block was the most frequently used (66/187, 35.3%) and replaced the standard general anesthetic for upper extremity surgery in 70 patients. The procedure success rate was 96.3% (180/187), and there were no observed patient complications.

Conclusions: This hybrid curriculum enabled the successful implementation of UGRA at a public hospital in Guatemala, safely expanding clinical capabilities and reducing reliance on general anesthesia for upper extremity surgery. This practical training model provides a framework for implementing UGRA in similar resource-limited hospitals.

背景:由于培训和设备方面的障碍,超声引导区域麻醉(UGRA)在低收入和中等收入国家(LMICs)仍未得到充分利用。便携式超声的最新进展(美国)和国际伙伴关系扩大了UGRA的可及性,提高了患者安全和护理质量。目的:本评价描述了危地马拉国立科泰佩克医院(HNC)麻醉师混合UGRA培训计划的发展和结果。方法:根据当地的需求和反馈,制定了UGRA的教育试点计划,包括四周的在线模块,一次面对面的教育会议和一个月的监督临床实践。评估遵循Kirkpatrick框架,采用全球区域麻醉课程参与(GRACE)模型的项目前和项目后调查。结果包括参与者的满意度、知识和技能的变化以及程序表现。比较培训前后的知识和技能评估,并记录10个月的临床数据。使用非参数测试来评估变化及其与性能结果的关联。结果:HNC的7名麻醉师均完成了培训计划。知识测试得分中位数百分比提高20.8% (5/24,IQR 13.5%-28.1%, r=0.899, P= 0.016),临床技能评分评分中位数百分比提高147.1% (1.8/5,IQR 96.9%-197.3%, r=0.904, P= 0.031), 4个月后提高131.4% (1.6/5,IQR 90.5%-172.3, r=0.909, P= 0.035)。参与者报告了高满意度和显著的感知改善和动力。第一个月在直接监督下进行了54次pnb,在10个月内记录了187个区块。锁骨上臂丛阻滞是最常用的(66.45%),在70例上肢手术中取代了标准全麻。手术成功率为96%(180/187),无并发症发生。结论:这一混合课程在危地马拉的一家公立医院成功实施了UGRA,安全地扩大了临床能力,减少了上肢手术对全身麻醉的依赖。这一实用的培训模式为在类似的资源有限的医院实施全民健康保险提供了一个框架。临床试验:
{"title":"Ultrasound-Guided Regional Anesthesia in a Resource-Limited Hospital: Prospective Pilot Study of a Hybrid Training Program.","authors":"Jakob E Gamboa, Inge Tamm-Daniels, Roland Flores, Nancy G Sarat Diaz, Mario A Villasenor, Mitchell A Gist, Aidan B Hoie, Christopher Kurinec, Colby G Simmons","doi":"10.2196/84181","DOIUrl":"10.2196/84181","url":null,"abstract":"<p><strong>Background: </strong>Ultrasound-guided regional anesthesia (UGRA) remains underused in low- and middle-income countries due to barriers to training and equipment. Recent advances in portable ultrasound devices and international partnerships have expanded access to UGRA, enhancing patient safety and quality of care.</p><p><strong>Objective: </strong>This study describes the development and outcomes of a hybrid UGRA training program for anesthesiologists at the Hospital Nacional de Coatepeque (HNC) in Guatemala.</p><p><strong>Methods: </strong>An educational pilot program for UGRA was developed based on local needs and feedback, comprising 4 weeks of online modules, an in-person educational conference, and 1 month of supervised clinical practice. Evaluation followed the Kirkpatrick framework using preprogram and postprogram surveys adapted from the Global Regional Anesthesia Curricular Engagement model. Outcomes included participants' satisfaction, change in knowledge and skill, and procedural performance. Knowledge and skill assessments were compared before and after the training, and clinical data were recorded for 10 months. Nonparametric tests were used to assess changes and associations with performance outcomes.</p><p><strong>Results: </strong>All 7 anesthesiologists at HNC completed the training program. Knowledge test scores improved by a median percentage increase of 20.8% (IQR 13.5%-28.1%; r=0.899; P=.02), and procedural skill rating scores increased by a median percentage of 147.1% (IQR 96.9%-197.3%; r=0.904; P=.03) at 1 month and 131.4% (IQR 90.5%-172.3%; r=0.909; P=.04) at 4 months after the program. Participants self-reported high satisfaction and substantial clinical improvement and motivation. A total of 54 peripheral nerve blocks were performed under direct supervision in the first month, with 187 blocks recorded over 10 months. The supraclavicular brachial plexus block was the most frequently used (66/187, 35.3%) and replaced the standard general anesthetic for upper extremity surgery in 70 patients. The procedure success rate was 96.3% (180/187), and there were no observed patient complications.</p><p><strong>Conclusions: </strong>This hybrid curriculum enabled the successful implementation of UGRA at a public hospital in Guatemala, safely expanding clinical capabilities and reducing reliance on general anesthesia for upper extremity surgery. This practical training model provides a framework for implementing UGRA in similar resource-limited hospitals.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":" ","pages":"e84181"},"PeriodicalIF":3.2,"publicationDate":"2026-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145811539","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
JMIR Medical Education
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1