首页 > 最新文献

JMIR Medical Education最新文献

英文 中文
Deconstructing Participant Behaviors in Virtual Reality Simulation: Ethnographic Analysis. 解构虚拟现实模拟中的参与者行为:民族志分析。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-27 DOI: 10.2196/65886
Daniel Loeb, Jamie Shoemaker, Kelly Ely, Matthew Zackoff
<p><strong>Background: </strong>Virtual reality (VR)-based simulation is an increasingly popular tool for simulation-based medical education, immersing participants in a realistic, 3D world where health care professionals can observe nuanced examination findings, such as subtle indicators of respiratory distress and skin perfusion. However, it remains unknown how the VR environment affects participant behavior and attention.</p><p><strong>Objective: </strong>This study aimed to describe clinician attention and decision-making behaviors during interprofessional pediatric resuscitation simulations performed in VR. We used video-based focused ethnography to describe how participant attention and behavior are altered in the VR environment and reflect how these changes may affect the educational profile of VR simulation.</p><p><strong>Methods: </strong>The research team analyzed scenarios with the question, "How does a completely virtual reality environment alter participant attention and behavior, and how might these changes impact educational goals?" Video-based focused ethnography consisting of data collection, analysis, and pattern explanation was conducted by experts in critical care, resuscitation, simulation, and medical education until data saturation was achieved.</p><p><strong>Results: </strong>Fifteen interprofessional VR simulation sessions featuring the same scenario-a child with pneumonia and sepsis-were evaluated. Three major themes emerged: Source of Truth, Cognitive Focus, and Fidelity Breakers. Source of Truth explores how participants gather and synthesize information in a VR environment. Participants used the patient's physical examination over ancillary data sources, such as the cardiorespiratory monitor, returning to the monitor when the examination did not align with expectations. Cognitive Focus describes the interplay between thinking, communicating, and doing during a VR simulation. The VR setting imposed unique cognitive demands, requiring participants to process information from multiple sources, make rapid decisions, and execute tasks during the scenario. Participants experienced increased task burden when virtual tasks did not mirror real-world procedures, leading to delays and fixation on certain actions. Fidelity Breakers reflects how technical and environmental factors disrupted focus and hindered learning. Navigational challenges, such as unintended teleportation and difficulties interacting with the virtual patient and equipment, disrupted participant immersion. These challenges underscore the current limitations of VR in reproducing the tactile and procedural aspects of real clinical care.</p><p><strong>Conclusions: </strong>Participants' focus on the physical examination findings in VR, as opposed to the cardiorespiratory monitor, potentially indicates simulation of an identical, more patient examination-centered approach to clinical data gathering. In addition, the multiple data sources allowed for participant cog
背景:基于虚拟现实(VR)的模拟是一种日益流行的基于模拟的医学教育工具,它使参与者沉浸在一个逼真的3D世界中,在这个世界中,卫生保健专业人员可以观察到细微的检查结果,例如呼吸窘迫和皮肤灌注的细微指标。然而,VR环境如何影响参与者的行为和注意力仍然是未知的。目的:本研究旨在描述临床医生在VR中进行跨专业儿科复苏模拟时的注意和决策行为。我们使用基于视频的重点人种学来描述参与者的注意力和行为在VR环境中是如何改变的,并反映这些变化如何影响VR模拟的教育概况。方法:研究小组分析了这样一个问题:“一个完全虚拟现实的环境是如何改变参与者的注意力和行为的,这些变化是如何影响教育目标的?”由重症监护、复苏、模拟和医学教育方面的专家进行数据收集、分析和模式解释,直至达到数据饱和。结果:评估了15个跨专业VR模拟会话,这些会话具有相同的场景-一个患有肺炎和败血症的儿童。三个主要的主题出现了:真相的来源,认知焦点和忠诚破坏者。真相之源探索参与者如何在虚拟现实环境中收集和综合信息。参与者使用患者的身体检查而不是辅助数据源,如心肺监护仪,当检查结果与预期不一致时返回监护仪。认知焦点描述了在VR模拟过程中思考、交流和行动之间的相互作用。VR设置施加了独特的认知需求,要求参与者处理来自多个来源的信息,快速做出决策,并在场景中执行任务。当虚拟任务不能反映真实世界的过程时,参与者的任务负担会增加,从而导致延迟和固定在某些动作上。“保真破坏者”反映了技术和环境因素是如何扰乱注意力和阻碍学习的。导航方面的挑战,如意外的瞬间移动和与虚拟病人和设备互动的困难,破坏了参与者的沉浸感。这些挑战强调了当前VR在再现真实临床护理的触觉和程序方面的局限性。结论:参与者在VR中对身体检查结果的关注,而不是心肺监护仪,可能表明模拟了一种相同的,更以患者检查为中心的临床数据收集方法。此外,多重数据来源允许参与者的认知负荷和任务负担,可能更好地反映现实生活中的临床护理。然而,需要偏离现实世界任务完成的技术功能,以及VR中的其他导航和互动挑战,导致保真度的破坏,并将注意力从学习目标转移。这些发现强调需要继续研究模拟模式、保真度和技术挑战如何影响参与者的注意力和行为,以便在期望的学习目标和训练模式之间进行深思熟虑的协调。
{"title":"Deconstructing Participant Behaviors in Virtual Reality Simulation: Ethnographic Analysis.","authors":"Daniel Loeb, Jamie Shoemaker, Kelly Ely, Matthew Zackoff","doi":"10.2196/65886","DOIUrl":"10.2196/65886","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;Virtual reality (VR)-based simulation is an increasingly popular tool for simulation-based medical education, immersing participants in a realistic, 3D world where health care professionals can observe nuanced examination findings, such as subtle indicators of respiratory distress and skin perfusion. However, it remains unknown how the VR environment affects participant behavior and attention.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Objective: &lt;/strong&gt;This study aimed to describe clinician attention and decision-making behaviors during interprofessional pediatric resuscitation simulations performed in VR. We used video-based focused ethnography to describe how participant attention and behavior are altered in the VR environment and reflect how these changes may affect the educational profile of VR simulation.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;The research team analyzed scenarios with the question, \"How does a completely virtual reality environment alter participant attention and behavior, and how might these changes impact educational goals?\" Video-based focused ethnography consisting of data collection, analysis, and pattern explanation was conducted by experts in critical care, resuscitation, simulation, and medical education until data saturation was achieved.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;Fifteen interprofessional VR simulation sessions featuring the same scenario-a child with pneumonia and sepsis-were evaluated. Three major themes emerged: Source of Truth, Cognitive Focus, and Fidelity Breakers. Source of Truth explores how participants gather and synthesize information in a VR environment. Participants used the patient's physical examination over ancillary data sources, such as the cardiorespiratory monitor, returning to the monitor when the examination did not align with expectations. Cognitive Focus describes the interplay between thinking, communicating, and doing during a VR simulation. The VR setting imposed unique cognitive demands, requiring participants to process information from multiple sources, make rapid decisions, and execute tasks during the scenario. Participants experienced increased task burden when virtual tasks did not mirror real-world procedures, leading to delays and fixation on certain actions. Fidelity Breakers reflects how technical and environmental factors disrupted focus and hindered learning. Navigational challenges, such as unintended teleportation and difficulties interacting with the virtual patient and equipment, disrupted participant immersion. These challenges underscore the current limitations of VR in reproducing the tactile and procedural aspects of real clinical care.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;Participants' focus on the physical examination findings in VR, as opposed to the cardiorespiratory monitor, potentially indicates simulation of an identical, more patient examination-centered approach to clinical data gathering. In addition, the multiple data sources allowed for participant cog","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e65886"},"PeriodicalIF":3.2,"publicationDate":"2025-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12571426/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145379136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Application of AI Communication Training Tools in Medical Undergraduate Education: Mixed Methods Feasibility Study Within a Primary Care Context. 人工智能交流培训工具在医学本科教育中的应用:初级保健背景下的混合方法可行性研究
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-24 DOI: 10.2196/70766
Chris Jacobs, Hans Johnson, Nina Tan, Kirsty Brownlie, Richard Joiner, Trevor Thompson
<p><strong>Background: </strong>Effective communication is fundamental to high-quality health care delivery, influencing patient satisfaction, adherence to treatment plans, and clinical outcomes. However, communication skills training for medical undergraduates often faces challenges in scalability, resource allocation, and personalization. Traditional methods, such as role-playing with standardized patients, are resource intensive and may not provide consistent feedback tailored to individual learners' needs. Artificial intelligence (AI) offers realistic patient interactions for education.</p><p><strong>Objective: </strong>This study aims to investigate the application of AI communication training tools in medical undergraduate education within a primary care context. The study evaluates the effectiveness, usability, and impact of AI virtual patients (VPs) on medical students' experience in communication skills practice.</p><p><strong>Methods: </strong>The study used a mixed methods sequential explanatory design, comprising a quantitative survey followed by qualitative focus group discussions. Eighteen participants, including 15 medical students and 3 practicing doctors, engaged with an AI VP simulating a primary care consultation for prostate cancer risk assessment. The AI VP was designed using a large language model and natural voice synthesis to create realistic patient interactions. The survey assessed 5 domains: fidelity, immersion, intrinsic motivation, debriefing, and system usability. Focus groups were used to explore participants' experiences, challenges, and perceived educational value of the AI tool.</p><p><strong>Results: </strong>Significant positive responses emerged against a neutral baseline, with the following median scores: intrinsic motivation 16.5 of 20.0 (IQR 15.0-18.0; d=2.09, P<.001), system usability 12.0 of 15.0 (IQR 11.5-12.5; d=2.18, P<.001), and psychological safety 5.0 of 5.0 (IQR 5.0-5.0; d=4.78, P<.001). Fidelity (median score 6.0/10.0, IQR 5.2-7.0; d=-0.08, P=.02) and immersion (median score 8.5/15.0, IQR 7.0-9.8; d=0.25 P=.08) were moderately rated. The overall Immersive Technology Evaluation Measure scores showed a high positive learning experience: median 47.5 of 65.0 (IQR 43.0-51.2; d=2.00, P<.001). Qualitative analysis identified 3 major themes across 11 subthemes, with participants highlighting both technical limitations and educational value. Participants valued the safe practice environment and the ability to receive immediate feedback.</p><p><strong>Conclusions: </strong>AI VP technology shows promising potential for communication skills training despite the current realism limitations. While it does not yet match human standardized patient authenticity, the technology has achieved sufficient fidelity to support meaningful educational interactions, and this study identified clear areas for improvement. The integration of AI into medical curricula represents a promising avenue for innovation in medical edu
背景:有效的沟通是高质量医疗服务的基础,影响患者满意度、对治疗计划的依从性和临床结果。然而,医学本科生沟通技能培训在可扩展性、资源分配和个性化等方面面临挑战。传统的方法,如对标准化患者进行角色扮演,是资源密集型的,可能无法提供针对个别学习者需求的一致反馈。人工智能(AI)为教育提供了现实的患者互动。目的:本研究旨在探讨人工智能沟通培训工具在初级保健背景下医学本科教育中的应用。本研究评估AI虚拟病人(VPs)对医学生沟通技巧实践体验的有效性、可用性和影响。方法:本研究采用混合方法序贯解释设计,包括定量调查和定性焦点小组讨论。18名参与者,包括15名医学生和3名执业医生,与人工智能副总裁进行模拟前列腺癌风险评估的初级保健咨询。人工智能副总裁使用大型语言模型和自然语音合成来设计,以创建逼真的患者交互。调查评估了5个领域:保真度、沉浸感、内在动机、汇报和系统可用性。焦点小组被用来探讨参与者的经历、挑战以及对人工智能工具的感知教育价值。结果:在中性基线下出现了显著的积极反应,其中位数得分为:内在动机16.5 (IQR为15.0-18.0;d=2.09, p)。结论:尽管目前现实主义的局限性,人工智能VP技术在沟通技能培训中显示出良好的潜力。虽然它还不能与人类标准化的患者真实性相匹配,但该技术已经达到了足够的保真度,可以支持有意义的教育互动,并且本研究确定了明确的改进领域。将人工智能纳入医学课程是医学教育创新的一个有希望的途径,有可能提高培训计划的质量和有效性。
{"title":"Application of AI Communication Training Tools in Medical Undergraduate Education: Mixed Methods Feasibility Study Within a Primary Care Context.","authors":"Chris Jacobs, Hans Johnson, Nina Tan, Kirsty Brownlie, Richard Joiner, Trevor Thompson","doi":"10.2196/70766","DOIUrl":"10.2196/70766","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;Effective communication is fundamental to high-quality health care delivery, influencing patient satisfaction, adherence to treatment plans, and clinical outcomes. However, communication skills training for medical undergraduates often faces challenges in scalability, resource allocation, and personalization. Traditional methods, such as role-playing with standardized patients, are resource intensive and may not provide consistent feedback tailored to individual learners' needs. Artificial intelligence (AI) offers realistic patient interactions for education.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Objective: &lt;/strong&gt;This study aims to investigate the application of AI communication training tools in medical undergraduate education within a primary care context. The study evaluates the effectiveness, usability, and impact of AI virtual patients (VPs) on medical students' experience in communication skills practice.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;The study used a mixed methods sequential explanatory design, comprising a quantitative survey followed by qualitative focus group discussions. Eighteen participants, including 15 medical students and 3 practicing doctors, engaged with an AI VP simulating a primary care consultation for prostate cancer risk assessment. The AI VP was designed using a large language model and natural voice synthesis to create realistic patient interactions. The survey assessed 5 domains: fidelity, immersion, intrinsic motivation, debriefing, and system usability. Focus groups were used to explore participants' experiences, challenges, and perceived educational value of the AI tool.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;Significant positive responses emerged against a neutral baseline, with the following median scores: intrinsic motivation 16.5 of 20.0 (IQR 15.0-18.0; d=2.09, P&lt;.001), system usability 12.0 of 15.0 (IQR 11.5-12.5; d=2.18, P&lt;.001), and psychological safety 5.0 of 5.0 (IQR 5.0-5.0; d=4.78, P&lt;.001). Fidelity (median score 6.0/10.0, IQR 5.2-7.0; d=-0.08, P=.02) and immersion (median score 8.5/15.0, IQR 7.0-9.8; d=0.25 P=.08) were moderately rated. The overall Immersive Technology Evaluation Measure scores showed a high positive learning experience: median 47.5 of 65.0 (IQR 43.0-51.2; d=2.00, P&lt;.001). Qualitative analysis identified 3 major themes across 11 subthemes, with participants highlighting both technical limitations and educational value. Participants valued the safe practice environment and the ability to receive immediate feedback.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;AI VP technology shows promising potential for communication skills training despite the current realism limitations. While it does not yet match human standardized patient authenticity, the technology has achieved sufficient fidelity to support meaningful educational interactions, and this study identified clear areas for improvement. The integration of AI into medical curricula represents a promising avenue for innovation in medical edu","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e70766"},"PeriodicalIF":3.2,"publicationDate":"2025-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12551969/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145369013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Applications, Challenges, and Prospects of Generative Artificial Intelligence Empowering Medical Education: Scoping Review. 生成式人工智能在医学教育中的应用、挑战和前景:范围审查。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-23 DOI: 10.2196/71125
Yuhang Lin, Zhiheng Luo, Zicheng Ye, Nuoxi Zhong, Lijian Zhao, Long Zhang, Xiaolan Li, Zetao Chen, Yijia Chen
<p><strong>Background: </strong>Nowadays, generative artificial intelligence (GAI) drives medical education toward enhanced intelligence, personalization, and interactivity. With its vast generative abilities and diverse applications, GAI redefines how educational resources are accessed, teaching methods are implemented, and assessments are conducted.</p><p><strong>Objective: </strong>This study aimed to review the current applications of GAI in medical education; analyze its opportunities and challenges; identify its strengths and potential issues in educational methods, assessments, and resources; and capture GAI's rapid evolution and multidimensional applications in medical education, thereby providing a theoretical foundation for future practice.</p><p><strong>Methods: </strong>This scoping review used PubMed, Web of Science, and Scopus to analyze literature from January 2023 to October 2024, focusing on GAI applications in medical education. Following PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) guidelines, 5991 articles were retrieved, with 1304 duplicates removed. The 2-stage screening (title or abstract and full-text review) excluded 4564 articles and a supplementary search included 8 articles, yielding 131 studies for final synthesis. We included (1) studies addressing GAI's applications, challenges, or future directions in medical education, (2) empirical research, systematic reviews, and meta-analyses, and (3) English-language articles. We excluded commentaries, editorials, viewpoints, perspectives, short reports, or communications with low levels of evidence, non-GAI technologies, and studies centered on other fields of medical education (eg, nursing). We integrated quantitative analysis of publication trends and Human Development Index (HDI) with thematic analysis of applications, technical limitations, and ethical implications.</p><p><strong>Results: </strong>Analysis of 131 articles revealed that 74.0% (n=97) originated from countries or regions with very high HDI, with the United States contributing the most (n=33); 14.5% (n=19) were from high HDI countries, 5.3% (n=7) from medium HDI countries, and 2.2% (n=3) from low HDI countries, with 3.8% (n=5) involving cross-HDI collaborations. ChatGPT was the most studied GAI model (n=119), followed by Gemini (n=22), Copilot (n=11), Claude (n=6), and LLaMA (n=4). Thematic analysis indicated that GAI applications in medical education mainly embody the diversification of educational methods, scientific evaluation of educational assessments, and dynamic optimization of educational resources. However, it also highlighted current limitations and potential future challenges, including insufficient scene adaptability, data quality and information bias, overreliance, and ethical controversies.</p><p><strong>Conclusions: </strong>GAI application in medical education exhibits significant regional disparities in development, and model r
背景:如今,生成式人工智能(GAI)推动医学教育朝着增强智能、个性化和交互性的方向发展。GAI凭借其强大的生成能力和多样化的应用,重新定义了教育资源的获取、教学方法的实施和评估的进行。目的:综述GAI在医学教育中的应用现状;分析机遇与挑战;确定其在教育方法、评估和资源方面的优势和潜在问题;捕捉GAI的快速演变和在医学教育中的多维应用,从而为未来的实践提供理论基础。方法:使用PubMed、Web of Science和Scopus对2023年1月至2024年10月的文献进行分析,重点关注GAI在医学教育中的应用。按照PRISMA-ScR(系统评价和荟萃分析扩展范围评价的首选报告项目)指南,检索了5991篇文章,删除了1304篇重复的文章。两阶段筛选(标题或摘要和全文综述)排除了4564篇文章,补充检索包括8篇文章,最终合成131篇研究。我们纳入了(1)关于GAI在医学教育中的应用、挑战或未来方向的研究,(2)实证研究、系统综述和荟萃分析,以及(3)英语文章。我们排除了评论、社论、观点、观点、短报告、低证据水平的通讯、非gai技术和以医学教育其他领域(如护理)为中心的研究。我们将出版物趋势和人类发展指数(HDI)的定量分析与应用、技术限制和伦理影响的专题分析结合起来。结果:131篇文献分析显示,74.0% (n=97)的文献来自HDI极高的国家或地区,其中美国贡献最多(n=33);14.5% (n=19)来自高HDI国家,5.3% (n=7)来自中等HDI国家,2.2% (n=3)来自低HDI国家,其中3.8% (n=5)涉及跨HDI合作。ChatGPT是研究最多的GAI模型(n=119),其次是Gemini (n=22)、Copilot (n=11)、Claude (n=6)和LLaMA (n=4)。专题分析表明,GAI在医学教育中的应用主要体现在教育方法的多样化、教育评价的科学化、教育资源的动态优化等方面。然而,它也强调了当前的局限性和潜在的未来挑战,包括场景适应性不足、数据质量和信息偏差、过度依赖以及伦理争议。结论:GAI在医学教育中的应用发展存在显著的地区差异,模型研究统计反映了研究者的一定使用偏好。GAI具有增强医学教育能力的潜力,但广泛采用GAI需要克服复杂的技术和伦理挑战。我们以共生代理理论为基础,主张建立资源-方法-评估三方模型,开发专业模型,构建综合大语言模型与专业模型相结合的综合体系,促进资源共享,完善伦理治理,构建人机共生的教育生态系统,实现科技-人文深度融合,推动医学教育向更高效率和以人为本的方向发展。
{"title":"Applications, Challenges, and Prospects of Generative Artificial Intelligence Empowering Medical Education: Scoping Review.","authors":"Yuhang Lin, Zhiheng Luo, Zicheng Ye, Nuoxi Zhong, Lijian Zhao, Long Zhang, Xiaolan Li, Zetao Chen, Yijia Chen","doi":"10.2196/71125","DOIUrl":"10.2196/71125","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;Nowadays, generative artificial intelligence (GAI) drives medical education toward enhanced intelligence, personalization, and interactivity. With its vast generative abilities and diverse applications, GAI redefines how educational resources are accessed, teaching methods are implemented, and assessments are conducted.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Objective: &lt;/strong&gt;This study aimed to review the current applications of GAI in medical education; analyze its opportunities and challenges; identify its strengths and potential issues in educational methods, assessments, and resources; and capture GAI's rapid evolution and multidimensional applications in medical education, thereby providing a theoretical foundation for future practice.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;This scoping review used PubMed, Web of Science, and Scopus to analyze literature from January 2023 to October 2024, focusing on GAI applications in medical education. Following PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) guidelines, 5991 articles were retrieved, with 1304 duplicates removed. The 2-stage screening (title or abstract and full-text review) excluded 4564 articles and a supplementary search included 8 articles, yielding 131 studies for final synthesis. We included (1) studies addressing GAI's applications, challenges, or future directions in medical education, (2) empirical research, systematic reviews, and meta-analyses, and (3) English-language articles. We excluded commentaries, editorials, viewpoints, perspectives, short reports, or communications with low levels of evidence, non-GAI technologies, and studies centered on other fields of medical education (eg, nursing). We integrated quantitative analysis of publication trends and Human Development Index (HDI) with thematic analysis of applications, technical limitations, and ethical implications.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;Analysis of 131 articles revealed that 74.0% (n=97) originated from countries or regions with very high HDI, with the United States contributing the most (n=33); 14.5% (n=19) were from high HDI countries, 5.3% (n=7) from medium HDI countries, and 2.2% (n=3) from low HDI countries, with 3.8% (n=5) involving cross-HDI collaborations. ChatGPT was the most studied GAI model (n=119), followed by Gemini (n=22), Copilot (n=11), Claude (n=6), and LLaMA (n=4). Thematic analysis indicated that GAI applications in medical education mainly embody the diversification of educational methods, scientific evaluation of educational assessments, and dynamic optimization of educational resources. However, it also highlighted current limitations and potential future challenges, including insufficient scene adaptability, data quality and information bias, overreliance, and ethical controversies.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;GAI application in medical education exhibits significant regional disparities in development, and model r","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e71125"},"PeriodicalIF":3.2,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12547994/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145348932","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated Evaluation of Reflection and Feedback Quality in Workplace-Based Assessments by Using Natural Language Processing: Cross-Sectional Competency-Based Medical Education Study. 基于自然语言处理的工作场所评估中反思和反馈质量的自动评估:基于能力的横断面医学教育研究。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-22 DOI: 10.2196/81718
Jeng-Wen Chen, Hai-Lun Tu, Chun-Hsiang Chang, Wei-Chung Hsu, Pa-Chun Wang, Chun-Hou Liao, Mingchih Chen

Background: Competency-based medical education relies heavily on high-quality narrative reflections and feedback within workplace-based assessments. However, evaluating these narratives at scale remains a significant challenge.

Objective: This study aims to develop and apply natural language processing (NLP) models to evaluate the quality of resident reflections and faculty feedback documented in Entrustable Professional Activities (EPAs) on Taiwan's nationwide Emyway platform for otolaryngology residency training.

Methods: This 4-year cross-sectional study analyzes 300 randomly sampled EPA assessments from 2021 to 2025, covering a pilot year and 3 full implementation years. Two medical education experts independently rated the narratives based on relevance, specificity, and the presence of reflective or improvement-focused language. Narratives were categorized into 4 quality levels-effective, moderate, ineffective, or irrelevant-and then dichotomized into high quality and low quality. We compared the performance of logistic regression, support vector machine, and bidirectional encoder representations from transformers (BERT) models in classifying narrative quality. The best performing model was then applied to track quality trends over time.

Results: The BERT model, a multilingual pretrained language model, outperformed other approaches, achieving 85% and 92% accuracy in binary classification for resident reflections and faculty feedback, respectively. The accuracy for the 4-level classification was 67% for both. Longitudinal analysis revealed significant increases in high-quality reflections (from 70.3% to 99.5%) and feedback (from 50.6% to 88.9%) over the study period.

Conclusions: BERT-based NLP demonstrated moderate-to-high accuracy in evaluating the narrative quality in EPA assessments, especially in the binary classification. While not a replacement for expert review, NLP models offer a valuable tool for monitoring narrative trends and enhancing formative feedback in competency-based medical education.

背景:基于能力的医学教育在很大程度上依赖于基于工作场所的评估中的高质量叙事反思和反馈。然而,大规模评估这些叙述仍然是一个重大挑战。摘要目的:本研究旨在建立自然语言处理(NLP)模型,以评估台湾Emyway耳鼻喉科住院医师培训平台“可信赖专业活动”(EPAs)中住院医师反思与教师反馈的品质。方法:这项为期4年的横断面研究分析了从2021年到2025年300个随机抽样的EPA评估,涵盖了一个试点年和3个全面实施年。两位医学教育专家根据相关性、特异性和反思性或以改进为重点的语言的存在对叙述进行了独立评级。叙述被分为4个质量水平——有效、中等、无效或不相关——然后被分为高质量和低质量。我们比较了逻辑回归、支持向量机和双向编码器转换器(BERT)模型在分类叙事质量方面的表现。然后将表现最好的模型应用于跟踪质量趋势。结果:BERT模型是一种多语言预训练的语言模型,在居民反映和教师反馈的二元分类中分别达到85%和92%的准确率,优于其他方法。4级分类的准确率均为67%。纵向分析显示,在研究期间,高质量的反射(从70.3%增加到99.5%)和反馈(从50.6%增加到88.9%)显著增加。结论:基于bert的自然语言处理在评价EPA的叙事质量方面具有中高的准确性,特别是在二元分类方面。虽然不能替代专家审查,但NLP模型为监测叙事趋势和增强基于能力的医学教育的形成性反馈提供了有价值的工具。
{"title":"Automated Evaluation of Reflection and Feedback Quality in Workplace-Based Assessments by Using Natural Language Processing: Cross-Sectional Competency-Based Medical Education Study.","authors":"Jeng-Wen Chen, Hai-Lun Tu, Chun-Hsiang Chang, Wei-Chung Hsu, Pa-Chun Wang, Chun-Hou Liao, Mingchih Chen","doi":"10.2196/81718","DOIUrl":"10.2196/81718","url":null,"abstract":"<p><strong>Background: </strong>Competency-based medical education relies heavily on high-quality narrative reflections and feedback within workplace-based assessments. However, evaluating these narratives at scale remains a significant challenge.</p><p><strong>Objective: </strong>This study aims to develop and apply natural language processing (NLP) models to evaluate the quality of resident reflections and faculty feedback documented in Entrustable Professional Activities (EPAs) on Taiwan's nationwide Emyway platform for otolaryngology residency training.</p><p><strong>Methods: </strong>This 4-year cross-sectional study analyzes 300 randomly sampled EPA assessments from 2021 to 2025, covering a pilot year and 3 full implementation years. Two medical education experts independently rated the narratives based on relevance, specificity, and the presence of reflective or improvement-focused language. Narratives were categorized into 4 quality levels-effective, moderate, ineffective, or irrelevant-and then dichotomized into high quality and low quality. We compared the performance of logistic regression, support vector machine, and bidirectional encoder representations from transformers (BERT) models in classifying narrative quality. The best performing model was then applied to track quality trends over time.</p><p><strong>Results: </strong>The BERT model, a multilingual pretrained language model, outperformed other approaches, achieving 85% and 92% accuracy in binary classification for resident reflections and faculty feedback, respectively. The accuracy for the 4-level classification was 67% for both. Longitudinal analysis revealed significant increases in high-quality reflections (from 70.3% to 99.5%) and feedback (from 50.6% to 88.9%) over the study period.</p><p><strong>Conclusions: </strong>BERT-based NLP demonstrated moderate-to-high accuracy in evaluating the narrative quality in EPA assessments, especially in the binary classification. While not a replacement for expert review, NLP models offer a valuable tool for monitoring narrative trends and enhancing formative feedback in competency-based medical education.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e81718"},"PeriodicalIF":3.2,"publicationDate":"2025-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12590046/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145348980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Insights Into History and Trends of Teaching and Learning in Stomatology Education: Bibliometric Analysis. 口腔医学教学的历史与趋势:文献计量学分析。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-20 DOI: 10.2196/66322
Ziang Zou, Linna Guo

Background: Stomatology education has experienced substantial transformations over recent decades. Nevertheless, a comprehensive summary encompassing the entirety of this field remains absent in the literature.

Objective: This study aimed to perform a bibliometric analysis to evaluate the research status, current focus, and emerging trends in this field over the last two decades.

Methods: We retrieved publications concerning teaching and learning in stomatology education from the Web of Science core collection covering the period from 2003 to 2023. Subsequently, we conducted a bibliometric analysis and visualization using R-Bibliometrix and CiteSpace.

Results: In total, 5528 publications focusing on teaching and learning in stomatology education were identified. The annual number of publications in this field has shown a consistent upward trend. The United States and the United Kingdom emerged as the leading contributors to research. Among academic institutions, the University of Iowa produced the highest number of publications. The Journal of Dental Education was identified as the journal with the highest citation. Wanchek T authored the most highly cited articles in the field. Emerging research hotspots were characterized by keywords such as "deep learning," "machine learning," "online learning," "virtual reality," and "convolutional neural network." The thematic map analysis further revealed that "surgery" and "accuracy" were categorized as emerging themes.

Conclusions: The visualization bibliometric analysis of the literature clearly depicts the current hotspots and emerging topics in stomatology education concerning teaching and learning. The findings are intended to serve as a reference to advance the development of stomatology education research globally.

背景:近几十年来,口腔医学教育经历了重大变革。然而,一个全面的总结,包括整个领域在文献中仍然缺席。目的:本研究旨在通过文献计量分析来评价近二十年来该领域的研究现状、研究热点和发展趋势。方法:检索Web of Science核心馆藏2003 - 2023年有关口腔医学教学的出版物。随后,我们使用R-Bibliometrix和CiteSpace进行了文献计量学分析和可视化。结果:共检索到口腔医学教学相关文献5528篇。这一领域的年度出版物数量呈现出持续上升的趋势。美国和英国成为研究的主要贡献者。在学术机构中,爱荷华大学发表的出版物数量最多。《牙科教育杂志》被确定为引用率最高的杂志。Wanchek T撰写了该领域被引用次数最多的文章。新兴研究热点以“深度学习”、“机器学习”、“在线学习”、“虚拟现实”和“卷积神经网络”等关键词为特征。专题地图分析进一步显示,“手术”和“准确性”被归类为新兴主题。结论:对文献进行可视化文献计量学分析,清晰地描绘了当前口腔医学教育中涉及教与学的热点和新兴课题。研究结果可为推动全球口腔医学教育研究的发展提供参考。
{"title":"Insights Into History and Trends of Teaching and Learning in Stomatology Education: Bibliometric Analysis.","authors":"Ziang Zou, Linna Guo","doi":"10.2196/66322","DOIUrl":"10.2196/66322","url":null,"abstract":"<p><strong>Background: </strong>Stomatology education has experienced substantial transformations over recent decades. Nevertheless, a comprehensive summary encompassing the entirety of this field remains absent in the literature.</p><p><strong>Objective: </strong>This study aimed to perform a bibliometric analysis to evaluate the research status, current focus, and emerging trends in this field over the last two decades.</p><p><strong>Methods: </strong>We retrieved publications concerning teaching and learning in stomatology education from the Web of Science core collection covering the period from 2003 to 2023. Subsequently, we conducted a bibliometric analysis and visualization using R-Bibliometrix and CiteSpace.</p><p><strong>Results: </strong>In total, 5528 publications focusing on teaching and learning in stomatology education were identified. The annual number of publications in this field has shown a consistent upward trend. The United States and the United Kingdom emerged as the leading contributors to research. Among academic institutions, the University of Iowa produced the highest number of publications. The Journal of Dental Education was identified as the journal with the highest citation. Wanchek T authored the most highly cited articles in the field. Emerging research hotspots were characterized by keywords such as \"deep learning,\" \"machine learning,\" \"online learning,\" \"virtual reality,\" and \"convolutional neural network.\" The thematic map analysis further revealed that \"surgery\" and \"accuracy\" were categorized as emerging themes.</p><p><strong>Conclusions: </strong>The visualization bibliometric analysis of the literature clearly depicts the current hotspots and emerging topics in stomatology education concerning teaching and learning. The findings are intended to serve as a reference to advance the development of stomatology education research globally.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e66322"},"PeriodicalIF":3.2,"publicationDate":"2025-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12536922/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145337635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correlation Between Electroencephalogram Brain-to-Brain Synchronization and Team Strategies and Tools to Enhance Performance and Patient Safety Scores During Online Hexad Virtual Simulation-Based Interprofessional Education: Cross-Sectional Correlational Study. 在基于在线十六进制虚拟模拟的跨专业教育中,脑电图脑对脑同步与团队策略和工具之间的相关性:横断面相关性研究。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-20 DOI: 10.2196/69725
Atthaphon Viriyopase, Khuansiri Narajeenron

Background: Team performance is crucial in crisis situations. Although the Thai version of Team Strategies and Tools to Enhance Performance and Patient Safety (TeamSTEPPS) has been validated, challenges remain due to its subjective evaluation. To date, no studies have examined the relationship between electroencephalogram (EEG) activity and team performance, as assessed by TeamSTEPPS, during virtual simulation-based interprofessional education (SIMBIE), where face-to-face communication is absent.

Objective: This study aims to investigate the correlation between EEG-based brain-to-brain synchronization and TeamSTEPPS scores in multiprofessional teams participating in virtual SIMBIE sessions.

Methods: This single-center study involved 90 participants (15 groups of 6 simulated professionals: 1 medical doctor, 2 nurses, 1 pharmacist, 1 medical technologist, and 1 radiological technologist). Each group completed two 30-minute virtual SIMBIE sessions focusing on team training in a crisis situation involving COVID-19 pneumonia with a difficult airway, resulting in 30 sessions in total. The TeamSTEPPS scores of each participant across 5 domains were independently assessed by 2 trained raters based on screen recording, and their average values were used. The scores of participants in the same session were aggregated to generate a group TeamSTEPPS score, representing group-level performance. EEG data were recorded using wireless EEG acquisition devices and computed for total interdependence (TI), which represents brain-to-brain synchronization. The TI values of participants in the same session were aggregated to produce a group TI, representing group-level brain-to-brain synchronization. We investigated the Pearson correlations between the TI and the scores at both the group and individual levels.

Results: Interrater reliability for the TeamSTEPPS scores among 12 raters indicated good agreement on average (mean 0.73, SD 0.18; range 0.32-0.999). At the individual level, the Pearson correlations between the TI and the scores were weak and not statistically significant across all TeamSTEPPS domains (all adjusted P≥.05). However, strongly negative, statistically significant correlations between the group TI and the group TeamSTEPPS scores in the alpha frequency band (8-12 Hz) of the anterior brain area were found across all TeamSTEPPS domains after correcting for multiple comparisons (mean -0.87, SD 0.06; range -0.93 to -0.8).

Conclusions: Strong negative correlations between the group TI and the group TeamSTEPPS scores were observed in the anterior alpha activity during online hexad virtual SIMBIE. These findings suggest that anterior alpha TI may serve as an objective metric for assessing TeamSTEPPS-based team performance.

背景:团队表现在危机情况下是至关重要的。尽管泰国版的团队战略和工具以提高绩效和患者安全(TeamSTEPPS)已经得到验证,但由于其主观评价,挑战仍然存在。到目前为止,还没有研究考察了在基于虚拟模拟的跨专业教育(SIMBIE)中脑电图(EEG)活动与团队绩效之间的关系,如TeamSTEPPS所评估的,在没有面对面交流的情况下。目的:研究参与SIMBIE虚拟游戏的多专业团队基于脑电图的脑对脑同步与TeamSTEPPS得分的相关性。方法:本研究采用单中心研究,共纳入90名受试者(15组,每组6名模拟专业人员:1名医生、2名护士、1名药剂师、1名医疗技师和1名放射技师)。每组完成两次30分钟的虚拟SIMBIE课程,重点是在COVID-19肺炎和气道困难的危机情况下进行团队训练,总共30次课程。每个参与者在5个领域的TeamSTEPPS得分由2名训练有素的评分员根据屏幕记录独立评估,并使用他们的平均值。同一会议中参与者的分数被汇总成一个小组TeamSTEPPS分数,代表小组水平的表现。利用无线脑电信号采集设备记录脑电图数据,并计算脑对脑同步的总依赖关系(TI)。同一会话中参与者的TI值被聚合以产生组TI,代表组级脑对脑同步。我们在小组和个人水平上调查了TI和分数之间的Pearson相关性。结果:12个评分者的TeamSTEPPS评分的信度平均一致性较好(平均值0.73,标准差0.18;范围0.32-0.999)。在个体水平上,TI与得分之间的Pearson相关性较弱,在所有TeamSTEPPS域中均无统计学意义(均校正P≥0.05)。然而,经过多次比较校正后,在所有TeamSTEPPS域中,TI组和TeamSTEPPS组在前脑区α频段(8-12 Hz)得分之间存在强烈的负相关,具有统计学意义(平均值-0.87,SD 0.06;范围-0.93至-0.8)。结论:TI组和TeamSTEPPS组在在线六回合虚拟SIMBIE的前α活动中观察到强烈的负相关。这些发现表明,前alpha TI可以作为评估基于teamsteps的团队绩效的客观指标。
{"title":"Correlation Between Electroencephalogram Brain-to-Brain Synchronization and Team Strategies and Tools to Enhance Performance and Patient Safety Scores During Online Hexad Virtual Simulation-Based Interprofessional Education: Cross-Sectional Correlational Study.","authors":"Atthaphon Viriyopase, Khuansiri Narajeenron","doi":"10.2196/69725","DOIUrl":"10.2196/69725","url":null,"abstract":"<p><strong>Background: </strong>Team performance is crucial in crisis situations. Although the Thai version of Team Strategies and Tools to Enhance Performance and Patient Safety (TeamSTEPPS) has been validated, challenges remain due to its subjective evaluation. To date, no studies have examined the relationship between electroencephalogram (EEG) activity and team performance, as assessed by TeamSTEPPS, during virtual simulation-based interprofessional education (SIMBIE), where face-to-face communication is absent.</p><p><strong>Objective: </strong>This study aims to investigate the correlation between EEG-based brain-to-brain synchronization and TeamSTEPPS scores in multiprofessional teams participating in virtual SIMBIE sessions.</p><p><strong>Methods: </strong>This single-center study involved 90 participants (15 groups of 6 simulated professionals: 1 medical doctor, 2 nurses, 1 pharmacist, 1 medical technologist, and 1 radiological technologist). Each group completed two 30-minute virtual SIMBIE sessions focusing on team training in a crisis situation involving COVID-19 pneumonia with a difficult airway, resulting in 30 sessions in total. The TeamSTEPPS scores of each participant across 5 domains were independently assessed by 2 trained raters based on screen recording, and their average values were used. The scores of participants in the same session were aggregated to generate a group TeamSTEPPS score, representing group-level performance. EEG data were recorded using wireless EEG acquisition devices and computed for total interdependence (TI), which represents brain-to-brain synchronization. The TI values of participants in the same session were aggregated to produce a group TI, representing group-level brain-to-brain synchronization. We investigated the Pearson correlations between the TI and the scores at both the group and individual levels.</p><p><strong>Results: </strong>Interrater reliability for the TeamSTEPPS scores among 12 raters indicated good agreement on average (mean 0.73, SD 0.18; range 0.32-0.999). At the individual level, the Pearson correlations between the TI and the scores were weak and not statistically significant across all TeamSTEPPS domains (all adjusted P≥.05). However, strongly negative, statistically significant correlations between the group TI and the group TeamSTEPPS scores in the alpha frequency band (8-12 Hz) of the anterior brain area were found across all TeamSTEPPS domains after correcting for multiple comparisons (mean -0.87, SD 0.06; range -0.93 to -0.8).</p><p><strong>Conclusions: </strong>Strong negative correlations between the group TI and the group TeamSTEPPS scores were observed in the anterior alpha activity during online hexad virtual SIMBIE. These findings suggest that anterior alpha TI may serve as an objective metric for assessing TeamSTEPPS-based team performance.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e69725"},"PeriodicalIF":3.2,"publicationDate":"2025-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12583944/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145337560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI's Accuracy in Extracting Learning Experiences From Clinical Practice Logs: Observational Study. 人工智能从临床实践日志中提取学习经验的准确性:观察性研究。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-15 DOI: 10.2196/68697
Takeshi Kondo, Hiroshi Nishigori

Background: Improving the quality of education in clinical settings requires an understanding of learners' experiences and learning processes. However, this is a significant burden on learners and educators. If learners' learning records could be automatically analyzed and their experiences could be visualized, this would enable real-time tracking of their progress. Large language models (LLMs) may be useful for this purpose, although their accuracy has not been sufficiently studied.

Objective: This study aimed to explore the accuracy of predicting the actual clinical experiences of medical students from their learning log data during clinical clerkship using LLMs.

Methods: This study was conducted at the Nagoya University School of Medicine. Learning log data from medical students participating in a clinical clerkship from April 22, 2024, to May 24, 2024, were used. The Model Core Curriculum for Medical Education was used as a template to extract experiences. OpenAI's ChatGPT was selected for this task after a comparison with other LLMs. Prompts were created using the learning log data and provided to ChatGPT to extract experiences, which were then listed. A web application using GPT-4-turbo was developed to automate this process. The accuracy of the extracted experiences was evaluated by comparing them with the corrected lists provided by the students.

Results: A total of 20 sixth-year medical students participated in this study, resulting in 40 datasets. The overall Jaccard index was 0.59 (95% CI 0.46-0.71), and the Cohen κ was 0.65 (95% CI 0.53-0.76). Overall sensitivity was 62.39% (95% CI 49.96%-74.81%), and specificity was 99.34% (95% CI 98.77%-99.92%). Category-specific performance varied: symptoms showed a sensitivity of 45.43% (95% CI 25.12%-65.75%) and specificity of 98.75% (95% CI 97.31%-100%), examinations showed a sensitivity of 46.76% (95% CI 25.67%-67.86%) and specificity of 98.84% (95% CI 97.81%-99.87%), and procedures achieved a sensitivity of 56.36% (95% CI 37.64%-75.08%) and specificity of 98.92% (95% CI 96.67%-100%). The results suggest that GPT-4-turbo accurately identified many of the actual experiences but missed some because of insufficient detail or a lack of student records.

Conclusions: This study demonstrated that LLMs such as GPT-4-turbo can predict clinical experiences from learning logs with high specificity but moderate sensitivity. Future improvements in AI models, providing feedback to medical students' learning logs and combining them with other data sources such as electronic medical records, may enhance the accuracy. Using artificial intelligence to analyze learning logs for assessment could reduce the burden on learners and educators while improving the quality of educational assessments in medical education.

背景:提高临床教育质量需要了解学习者的经历和学习过程。然而,这对学习者和教育者来说是一个沉重的负担。如果学习者的学习记录可以自动分析,他们的经验可以可视化,这将使实时跟踪他们的进步成为可能。大型语言模型(llm)可能对这一目的有用,尽管它们的准确性还没有得到充分的研究。目的:本研究旨在探讨利用LLMs从医学生临床见习学习日志数据预测医学生实际临床经验的准确性。方法:本研究在名古屋大学医学院进行。使用2024年4月22日至2024年5月24日参加临床实习的医学生的学习日志数据。以《医学教育核心课程示范》为模板提取经验。经过与其他llm的比较,我们选择OpenAI的ChatGPT来完成这个任务。使用学习日志数据创建提示,并提供给ChatGPT以提取经验,然后将其列出。开发了一个使用GPT-4-turbo的web应用程序来自动化此过程。通过将所提取的经验与学生提供的更正列表进行比较,来评估其准确性。结果:共有20名六年级医学生参与本研究,共获得40个数据集。总体Jaccard指数为0.59 (95% CI 0.46 ~ 0.71), Cohen κ为0.65 (95% CI 0.53 ~ 0.76)。总敏感性为62.39% (95% CI 49.96% ~ 74.81%),特异性为99.34% (95% CI 98.77% ~ 99.92%)。分类特异性表现不同:症状的敏感性为45.43% (95% CI 25.12%-65.75%),特异性为98.75% (95% CI 97.31%-100%),检查的敏感性为46.76% (95% CI 25.67%-67.86%),特异性为98.84% (95% CI 97.81%-99.87%),手术的敏感性为56.36% (95% CI 37.64%-75.08%),特异性为98.92% (95% CI 96.67%-100%)。结果表明,GPT-4-turbo准确地识别了许多实际经历,但由于细节不足或缺乏学生记录而遗漏了一些。结论:本研究表明,GPT-4-turbo等LLMs可以通过学习日志预测临床经验,特异性高,敏感性中等。AI模型的未来改进,为医学生的学习日志提供反馈,并将其与电子病历等其他数据源相结合,可能会提高准确性。利用人工智能分析学习日志进行评估可以减轻学习者和教育者的负担,同时提高医学教育中教育评估的质量。
{"title":"AI's Accuracy in Extracting Learning Experiences From Clinical Practice Logs: Observational Study.","authors":"Takeshi Kondo, Hiroshi Nishigori","doi":"10.2196/68697","DOIUrl":"10.2196/68697","url":null,"abstract":"<p><strong>Background: </strong>Improving the quality of education in clinical settings requires an understanding of learners' experiences and learning processes. However, this is a significant burden on learners and educators. If learners' learning records could be automatically analyzed and their experiences could be visualized, this would enable real-time tracking of their progress. Large language models (LLMs) may be useful for this purpose, although their accuracy has not been sufficiently studied.</p><p><strong>Objective: </strong>This study aimed to explore the accuracy of predicting the actual clinical experiences of medical students from their learning log data during clinical clerkship using LLMs.</p><p><strong>Methods: </strong>This study was conducted at the Nagoya University School of Medicine. Learning log data from medical students participating in a clinical clerkship from April 22, 2024, to May 24, 2024, were used. The Model Core Curriculum for Medical Education was used as a template to extract experiences. OpenAI's ChatGPT was selected for this task after a comparison with other LLMs. Prompts were created using the learning log data and provided to ChatGPT to extract experiences, which were then listed. A web application using GPT-4-turbo was developed to automate this process. The accuracy of the extracted experiences was evaluated by comparing them with the corrected lists provided by the students.</p><p><strong>Results: </strong>A total of 20 sixth-year medical students participated in this study, resulting in 40 datasets. The overall Jaccard index was 0.59 (95% CI 0.46-0.71), and the Cohen κ was 0.65 (95% CI 0.53-0.76). Overall sensitivity was 62.39% (95% CI 49.96%-74.81%), and specificity was 99.34% (95% CI 98.77%-99.92%). Category-specific performance varied: symptoms showed a sensitivity of 45.43% (95% CI 25.12%-65.75%) and specificity of 98.75% (95% CI 97.31%-100%), examinations showed a sensitivity of 46.76% (95% CI 25.67%-67.86%) and specificity of 98.84% (95% CI 97.81%-99.87%), and procedures achieved a sensitivity of 56.36% (95% CI 37.64%-75.08%) and specificity of 98.92% (95% CI 96.67%-100%). The results suggest that GPT-4-turbo accurately identified many of the actual experiences but missed some because of insufficient detail or a lack of student records.</p><p><strong>Conclusions: </strong>This study demonstrated that LLMs such as GPT-4-turbo can predict clinical experiences from learning logs with high specificity but moderate sensitivity. Future improvements in AI models, providing feedback to medical students' learning logs and combining them with other data sources such as electronic medical records, may enhance the accuracy. Using artificial intelligence to analyze learning logs for assessment could reduce the burden on learners and educators while improving the quality of educational assessments in medical education.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e68697"},"PeriodicalIF":3.2,"publicationDate":"2025-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12529426/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145303860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From Hype to Implementation: Embedding GPT-4o in Medical Education. 从炒作到实施:将gpt - 40嵌入医学教育。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-15 DOI: 10.2196/79309
Sumaia Sabouni, Mohammad-Adel Moufti, Mohamed Hassan Taha

Unlabelled: The release of GPT-4 Omni (GPT-4o), an advanced multimodal generative artificial intelligence (AI) model, generated substantial enthusiasm in the field of higher education. However, one year later, medical education continues to face significant challenges, demonstrating the need to move from initial experimentation with the integration of multimodal AIs in medical education toward meaningful integration. In this Viewpoint, we argue that GPT-4o's true value lies not in novelty, but in its potential to enhance training in communication skills, clinical reasoning, and procedural skills by offering real-time simulations and adaptive learning experiences using text, audio, and visual inputs in a safe, immersive, and cost-effective environment. We explore how this innovation has made it possible to address key medical educational challenges by simulating realistic patient interactions, offering personalized feedback, and reducing educator workloads and costs, where traditional teaching methods struggle to replicate the complexity and dynamism of real-world clinical scenarios. However, we also address the critical challenges of this approach, including data accuracy, bias, and ethical decision-making. Rather than seeing GPT-4o as a replacement, we propose its use as a strategic supplement, scaffolded into curriculum frameworks and evaluated through ongoing research. As the focus shifts from AI novelty to sustainable implementation, we call on educators, policymakers, and curriculum designers to establish governance mechanisms, pilot evaluation strategies, and develop faculty training. The future of AI in medical education depends not on the next breakthrough, but on how we integrate today's tools with intention and rigor.

未标记:GPT-4 Omni (gpt - 40)的发布,一种先进的多模态生成人工智能(AI)模型,在高等教育领域引起了极大的热情。然而,一年后,医学教育继续面临重大挑战,这表明需要从最初的实验转向医学教育中多模式人工智能的整合,以实现有意义的整合。在这一观点中,我们认为gpt - 40的真正价值不在于新颖,而在于它通过提供实时模拟和自适应学习体验,在安全、沉浸式和经济高效的环境中使用文本、音频和视觉输入,增强沟通技巧、临床推理和程序技能的培训潜力。在传统教学方法难以复制现实世界临床场景的复杂性和动态性的情况下,我们探索这种创新如何通过模拟真实的患者互动、提供个性化反馈、减少教育者的工作量和成本,使解决关键的医学教育挑战成为可能。然而,我们也解决了这种方法的关键挑战,包括数据准确性、偏见和道德决策。与其将gpt - 40视为替代品,我们建议将其作为一种战略补充,纳入课程框架,并通过正在进行的研究进行评估。随着焦点从人工智能的新颖性转向可持续实施,我们呼吁教育工作者、政策制定者和课程设计师建立治理机制,试点评估策略,并开展教师培训。人工智能在医学教育中的未来不取决于下一个突破,而取决于我们如何将今天的工具与意图和严谨结合起来。
{"title":"From Hype to Implementation: Embedding GPT-4o in Medical Education.","authors":"Sumaia Sabouni, Mohammad-Adel Moufti, Mohamed Hassan Taha","doi":"10.2196/79309","DOIUrl":"10.2196/79309","url":null,"abstract":"<p><strong>Unlabelled: </strong>The release of GPT-4 Omni (GPT-4o), an advanced multimodal generative artificial intelligence (AI) model, generated substantial enthusiasm in the field of higher education. However, one year later, medical education continues to face significant challenges, demonstrating the need to move from initial experimentation with the integration of multimodal AIs in medical education toward meaningful integration. In this Viewpoint, we argue that GPT-4o's true value lies not in novelty, but in its potential to enhance training in communication skills, clinical reasoning, and procedural skills by offering real-time simulations and adaptive learning experiences using text, audio, and visual inputs in a safe, immersive, and cost-effective environment. We explore how this innovation has made it possible to address key medical educational challenges by simulating realistic patient interactions, offering personalized feedback, and reducing educator workloads and costs, where traditional teaching methods struggle to replicate the complexity and dynamism of real-world clinical scenarios. However, we also address the critical challenges of this approach, including data accuracy, bias, and ethical decision-making. Rather than seeing GPT-4o as a replacement, we propose its use as a strategic supplement, scaffolded into curriculum frameworks and evaluated through ongoing research. As the focus shifts from AI novelty to sustainable implementation, we call on educators, policymakers, and curriculum designers to establish governance mechanisms, pilot evaluation strategies, and develop faculty training. The future of AI in medical education depends not on the next breakthrough, but on how we integrate today's tools with intention and rigor.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e79309"},"PeriodicalIF":3.2,"publicationDate":"2025-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12527310/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145303795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Training Gaps in Digital Skills for the Cancer Health Care Workforce Based on Insights From Clinical Professionals, Nonclinical Professionals, and Patients and Caregivers: Qualitative Study. 基于临床专业人员、非临床专业人员、患者和护理人员见解的癌症卫生保健人员数字技能培训差距:定性研究。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-08 DOI: 10.2196/78490
David Liñares, Theologia Tsitsi, Noemí López-Rey, Wilfredo Guanipa-Sierra, Susana Aldecoa-Landesa, Carme Carrión, Daniela Cabutto, Deborah Moreno-Alonso, Clara Madrid-Alejos, Andreas Charalambous, Ana Clavería

Background: The integration of digital technologies is becoming increasingly essential in cancer care. However, limited digital health literacy among clinical and nonclinical cancer health care professionals poses significant challenges to effective implementation and sustainability over time. To address this, the European Union is prioritizing the development of targeted digital skills training programs for cancer care providers, the TRANSiTION project among them. A crucial initial step in this effort is conducting a comprehensive gap analysis to identify specific training needs.

Objective: The aim of this work is to identify training gaps and prioritize the digital skill development needs in the oncology health care workforce.

Methods: An importance-performance analysis (IPA) was conducted following a survey that assessed the performance and importance of 7 digital skills: information, communication, content creation, safety, eHealth problem-solving, ethics, and patient empowerment.

Results: A total of 67 participants from 11 European countries completed the study: 38 clinical professionals (CP), 16 nonclinical professionals (NCP), and 13 patients or caregivers (PC). CP acknowledged the need for a comprehensive training program that includes all 7 digital skills. Digital patient empowerment and safety skills emerge as the highest priorities for both CP and NCP. Conversely, NCP assigned a lower priority to digital content creation skills, and PC assigned a lower priority to digital information and ethical skills. The IPA also revealed discrepancies in digital communication skills across groups (H=6.50; P=.04).

Conclusions: The study showcased the pressing need for comprehensive digital skill training for cancer health care professionals across diverse backgrounds and health care systems in Europe, tailored to their occupation and care setting. Incorporating PC perspectives ensures a balanced approach to addressing these training gaps. These findings provide a valuable knowledge base for designing digital skills training programs, promoting a holistic approach that integrates the perspectives of the various stakeholders involved in digital cancer care.

背景:数字技术的整合在癌症治疗中变得越来越重要。然而,随着时间的推移,临床和非临床癌症医疗保健专业人员的数字健康素养有限,对有效实施和可持续性构成重大挑战。为了解决这个问题,欧盟正在优先为癌症护理提供者开发有针对性的数字技能培训计划,其中包括过渡项目。这项工作的关键初步步骤是进行全面的差距分析,以确定具体的培训需要。目的:这项工作的目的是确定培训差距,并优先考虑肿瘤卫生保健工作人员的数字技能发展需求。方法:通过对7项数字技能(信息、沟通、内容创作、安全、电子健康问题解决、伦理和患者授权)的绩效和重要性进行调查,进行重要性-绩效分析(IPA)。结果:来自11个欧洲国家的67名参与者完成了这项研究:38名临床专业人员(CP), 16名非临床专业人员(NCP)和13名患者或护理人员(PC)。CP承认需要一个全面的培训计划,包括所有7项数字技能。数字化患者赋权和安全技能成为新型冠状病毒和新型冠状病毒的最高优先事项。相反,NCP对数字内容创造技能的优先级较低,PC对数字信息和道德技能的优先级较低。IPA还揭示了不同群体在数字沟通技能方面的差异(H=6.50; P= 0.04)。结论:该研究表明,迫切需要针对欧洲不同背景和医疗保健系统的癌症医疗保健专业人员进行全面的数字技能培训,以适应他们的职业和护理环境。结合PC的观点确保了平衡的方法来解决这些培训差距。这些发现为设计数字技能培训计划提供了宝贵的知识基础,促进了一种整合数字化癌症治疗中各个利益相关者观点的整体方法。
{"title":"Training Gaps in Digital Skills for the Cancer Health Care Workforce Based on Insights From Clinical Professionals, Nonclinical Professionals, and Patients and Caregivers: Qualitative Study.","authors":"David Liñares, Theologia Tsitsi, Noemí López-Rey, Wilfredo Guanipa-Sierra, Susana Aldecoa-Landesa, Carme Carrión, Daniela Cabutto, Deborah Moreno-Alonso, Clara Madrid-Alejos, Andreas Charalambous, Ana Clavería","doi":"10.2196/78490","DOIUrl":"10.2196/78490","url":null,"abstract":"<p><strong>Background: </strong>The integration of digital technologies is becoming increasingly essential in cancer care. However, limited digital health literacy among clinical and nonclinical cancer health care professionals poses significant challenges to effective implementation and sustainability over time. To address this, the European Union is prioritizing the development of targeted digital skills training programs for cancer care providers, the TRANSiTION project among them. A crucial initial step in this effort is conducting a comprehensive gap analysis to identify specific training needs.</p><p><strong>Objective: </strong>The aim of this work is to identify training gaps and prioritize the digital skill development needs in the oncology health care workforce.</p><p><strong>Methods: </strong>An importance-performance analysis (IPA) was conducted following a survey that assessed the performance and importance of 7 digital skills: information, communication, content creation, safety, eHealth problem-solving, ethics, and patient empowerment.</p><p><strong>Results: </strong>A total of 67 participants from 11 European countries completed the study: 38 clinical professionals (CP), 16 nonclinical professionals (NCP), and 13 patients or caregivers (PC). CP acknowledged the need for a comprehensive training program that includes all 7 digital skills. Digital patient empowerment and safety skills emerge as the highest priorities for both CP and NCP. Conversely, NCP assigned a lower priority to digital content creation skills, and PC assigned a lower priority to digital information and ethical skills. The IPA also revealed discrepancies in digital communication skills across groups (H=6.50; P=.04).</p><p><strong>Conclusions: </strong>The study showcased the pressing need for comprehensive digital skill training for cancer health care professionals across diverse backgrounds and health care systems in Europe, tailored to their occupation and care setting. Incorporating PC perspectives ensures a balanced approach to addressing these training gaps. These findings provide a valuable knowledge base for designing digital skills training programs, promoting a holistic approach that integrates the perspectives of the various stakeholders involved in digital cancer care.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e78490"},"PeriodicalIF":3.2,"publicationDate":"2025-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12547342/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145253056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ChatGPT in Medical Education: Bibliometric and Visual Analysis. ChatGPT在医学教育中的应用:文献计量学和视觉分析。
IF 3.2 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2025-10-07 DOI: 10.2196/72356
Yuning Zhang, Xiaolu Xie, Qi Xu
<p><strong>Background: </strong>ChatGPT is a generative artificial intelligence-based chatbot developed by OpenAI. Since its release in the second half of 2022, it has been widely applied across various fields. In particular, the application of ChatGPT in medical education has become a significant trend. To gain a comprehensive understanding of the research developments and trends regarding ChatGPT in medical education, we conducted an extensive review and analysis of the current state of research in this field.</p><p><strong>Objective: </strong>This study used bibliometric and visualization analysis to explore the current state of research and development trends regarding ChatGPT in medical education.</p><p><strong>Methods: </strong>A bibliometric analysis of 407 articles on ChatGPT in medical education published between March 2023 and June 2025 was conducted using CiteSpace, VOSviewer, and Bibliometrix (RTool of RStudio). Visualization of countries, institutions, journals, authors, keywords, and references was also conducted.</p><p><strong>Results: </strong>This bibliometric analysis included a total of 407 studies. Research in this field began in 2023, showing a notable surge in annual publications until June 2025. The United States, China, Türkiye, the United Kingdom, and Canada produced the most publications. Networks of collaboration also formed among institutions. The University of California system was a core research institution, with 3.4% (14/407) of the publications and 0.17 betweenness centrality. BMC Medical Education, Medical Teacher, and the Journal of Medical Internet Research were all among the top 10 journals in terms of both publication volume and citation frequency. The most prolific author was Yavuz Selim Kiyak, who has established a stable collaboration network with Isil Irem Budakoglu and Ozlem Coskun. Author collaboration in this field is usually limited, with most academic research conducted by independent teams and little communication between teams. The most frequent keywords were "AI," "ChatGPT," and "medical education." Keyword analysis further revealed "educational assessment," "exam," and "clinical practice" as current research hot spots. The most cited paper was "Performance of ChatGPT on USMLE: Potential for AI-Assisted Medical Education Using Large Language Models," and the paper with the strongest citation burst was "Are ChatGPT's Knowledge and Interpretation Ability Comparable to Those of Medical Students in Korea for Taking a Parasitology Examination?: A Descriptive Study." Both papers focus on evaluating ChatGPT's performance in medical exams.</p><p><strong>Conclusions: </strong>This study reveals the significant potential of ChatGPT in medical education. As the technology improves, its applications will expand into more fields. To promote the diversification and effectiveness of ChatGPT in medical education, future research should strengthen interregional collaboration and enhance research quality. These fin
背景:ChatGPT是OpenAI公司开发的基于生成式人工智能的聊天机器人。自2022年下半年发布以来,它已被广泛应用于各个领域。特别是ChatGPT在医学教育中的应用已经成为一个重要的趋势。为了全面了解ChatGPT在医学教育中的研究进展和趋势,我们对该领域的研究现状进行了广泛的回顾和分析。目的:采用文献计量学和可视化分析方法,探讨ChatGPT在医学教育中的研究现状和发展趋势。方法:采用CiteSpace、VOSviewer和Bibliometrix (RStudio的RTool)对2023年3月至2025年6月发表的407篇医学教育领域ChatGPT相关文献进行文献计量学分析。还对国家、机构、期刊、作者、关键词和参考文献进行了可视化。结果:文献计量学分析共纳入407项研究。这一领域的研究始于2023年,直到2025年6月,年度出版物都出现了显著的增长。美国、中国、日本、英国和加拿大发表的出版物最多。各机构之间也形成了合作网络。加州大学系统是核心研究机构,发表量占3.4%(14/407),中间中心性为0.17。《BMC Medical Education》、《Medical Teacher》、《Journal of Medical Internet Research》均在发表量和被引频次排名前十的期刊之列。最多产的作者是Yavuz Selim Kiyak,他与Isil Irem Budakoglu和Ozlem Coskun建立了稳定的合作网络。该领域的作者合作通常是有限的,大多数学术研究都是由独立的团队进行的,团队之间的交流很少。最常见的关键词是“人工智能”、“ChatGPT”和“医学教育”。关键词分析进一步揭示了“教育评价”、“考试”和“临床实践”是当前的研究热点。被引次数最多的论文是《ChatGPT在USMLE上的表现:使用大型语言模型的人工智能辅助医学教育的潜力》,被引次数最多的论文是《ChatGPT的知识和解释能力与韩国医学生参加寄生虫学考试的能力是否相当?》:一项描述性研究。两篇论文都着重于评估ChatGPT在医学考试中的表现。结论:本研究揭示了ChatGPT在医学教育中的巨大潜力。随着技术的进步,它的应用将扩展到更多的领域。为了促进ChatGPT在医学教育中的多样化和有效性,未来的研究应加强区域间的合作,提高研究质量。这些发现为研究人员确定研究视角和指导未来的研究方向提供了有价值的见解。
{"title":"ChatGPT in Medical Education: Bibliometric and Visual Analysis.","authors":"Yuning Zhang, Xiaolu Xie, Qi Xu","doi":"10.2196/72356","DOIUrl":"10.2196/72356","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;ChatGPT is a generative artificial intelligence-based chatbot developed by OpenAI. Since its release in the second half of 2022, it has been widely applied across various fields. In particular, the application of ChatGPT in medical education has become a significant trend. To gain a comprehensive understanding of the research developments and trends regarding ChatGPT in medical education, we conducted an extensive review and analysis of the current state of research in this field.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Objective: &lt;/strong&gt;This study used bibliometric and visualization analysis to explore the current state of research and development trends regarding ChatGPT in medical education.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;A bibliometric analysis of 407 articles on ChatGPT in medical education published between March 2023 and June 2025 was conducted using CiteSpace, VOSviewer, and Bibliometrix (RTool of RStudio). Visualization of countries, institutions, journals, authors, keywords, and references was also conducted.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;This bibliometric analysis included a total of 407 studies. Research in this field began in 2023, showing a notable surge in annual publications until June 2025. The United States, China, Türkiye, the United Kingdom, and Canada produced the most publications. Networks of collaboration also formed among institutions. The University of California system was a core research institution, with 3.4% (14/407) of the publications and 0.17 betweenness centrality. BMC Medical Education, Medical Teacher, and the Journal of Medical Internet Research were all among the top 10 journals in terms of both publication volume and citation frequency. The most prolific author was Yavuz Selim Kiyak, who has established a stable collaboration network with Isil Irem Budakoglu and Ozlem Coskun. Author collaboration in this field is usually limited, with most academic research conducted by independent teams and little communication between teams. The most frequent keywords were \"AI,\" \"ChatGPT,\" and \"medical education.\" Keyword analysis further revealed \"educational assessment,\" \"exam,\" and \"clinical practice\" as current research hot spots. The most cited paper was \"Performance of ChatGPT on USMLE: Potential for AI-Assisted Medical Education Using Large Language Models,\" and the paper with the strongest citation burst was \"Are ChatGPT's Knowledge and Interpretation Ability Comparable to Those of Medical Students in Korea for Taking a Parasitology Examination?: A Descriptive Study.\" Both papers focus on evaluating ChatGPT's performance in medical exams.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;This study reveals the significant potential of ChatGPT in medical education. As the technology improves, its applications will expand into more fields. To promote the diversification and effectiveness of ChatGPT in medical education, future research should strengthen interregional collaboration and enhance research quality. These fin","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e72356"},"PeriodicalIF":3.2,"publicationDate":"2025-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12503443/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145245498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
JMIR Medical Education
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1