Yuki Kataoka, Ryuhei So, Masahiro Banno, Yasushi Tsujimoto
Background: Research capacity building (RCB) among health care professionals remains limited, particularly for those working outside academic institutions. Japan is experiencing a decline in original clinical research due to insufficient RCB infrastructure. Our previous hospital-based workshops were effective but faced geographical and sustainability constraints. We developed a fully online Scientific Research Works Peer Support Group (SRWS-PSG) model that addresses geographical and time-bound constraints and establishes a sustainable economic model. Mentees use online materials, receive support from mentors via a communication platform after formulating their research question, and transition into mentors upon publication.
Objective: We evaluated whether our model's theoretical benefits translated into actual program effectiveness in RCB among health care professionals.
Methods: We conducted a retrospective cohort study of health care professionals who participated in the SRWS-PSG program between September 2019 and January 2025. Mentees progressed through a structured modular curriculum covering systematic review methodology, from protocol development to manuscript preparation, with personalized mentoring support. We evaluated manuscript submission, program discontinuation, promotion to a mentor status, and mentor response time. We collected data from program records and chat logs. Manuscript submission was defined as mentor-confirmed submission of a systematic review manuscript to a peer-reviewed journal. Program discontinuation referred to formal withdrawal before manuscript submission. Mentor promotion was defined as acceptance of an invitation to serve as a junior mentor after manuscript submission. Mentor response time was the elapsed time from a mentee's question in the chat to the first reply by an assigned mentor.
Results: Of 85 mentees analyzed, 31 (36.5%) held academic degrees (PhD or MPH), and 68 (80%) were medical doctors. During a median follow-up of 10 months, 51 (60%) submitted manuscripts and 46 (90%) became mentors. Ten mentees (12%) discontinued the program. The median mentor response time was 0.8 hours, with 90% responding within 24 hours.
Conclusions: A majority of participants of SRWS-PSG submitted manuscripts. This fully online RCB program might address geographical barriers and provides an adaptable approach for RCB across diverse health care contexts.
{"title":"Effectiveness of a Fully Online Scientific Research Works Peer Support Group Model for Research Capacity Building Through Conducting Systematic Reviews Among Health Care Professionals: Retrospective Cohort Studies.","authors":"Yuki Kataoka, Ryuhei So, Masahiro Banno, Yasushi Tsujimoto","doi":"10.2196/78862","DOIUrl":"10.2196/78862","url":null,"abstract":"<p><strong>Background: </strong>Research capacity building (RCB) among health care professionals remains limited, particularly for those working outside academic institutions. Japan is experiencing a decline in original clinical research due to insufficient RCB infrastructure. Our previous hospital-based workshops were effective but faced geographical and sustainability constraints. We developed a fully online Scientific Research Works Peer Support Group (SRWS-PSG) model that addresses geographical and time-bound constraints and establishes a sustainable economic model. Mentees use online materials, receive support from mentors via a communication platform after formulating their research question, and transition into mentors upon publication.</p><p><strong>Objective: </strong>We evaluated whether our model's theoretical benefits translated into actual program effectiveness in RCB among health care professionals.</p><p><strong>Methods: </strong>We conducted a retrospective cohort study of health care professionals who participated in the SRWS-PSG program between September 2019 and January 2025. Mentees progressed through a structured modular curriculum covering systematic review methodology, from protocol development to manuscript preparation, with personalized mentoring support. We evaluated manuscript submission, program discontinuation, promotion to a mentor status, and mentor response time. We collected data from program records and chat logs. Manuscript submission was defined as mentor-confirmed submission of a systematic review manuscript to a peer-reviewed journal. Program discontinuation referred to formal withdrawal before manuscript submission. Mentor promotion was defined as acceptance of an invitation to serve as a junior mentor after manuscript submission. Mentor response time was the elapsed time from a mentee's question in the chat to the first reply by an assigned mentor.</p><p><strong>Results: </strong>Of 85 mentees analyzed, 31 (36.5%) held academic degrees (PhD or MPH), and 68 (80%) were medical doctors. During a median follow-up of 10 months, 51 (60%) submitted manuscripts and 46 (90%) became mentors. Ten mentees (12%) discontinued the program. The median mentor response time was 0.8 hours, with 90% responding within 24 hours.</p><p><strong>Conclusions: </strong>A majority of participants of SRWS-PSG submitted manuscripts. This fully online RCB program might address geographical barriers and provides an adaptable approach for RCB across diverse health care contexts.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e78862"},"PeriodicalIF":3.2,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12490813/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145565822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Unlabelled: The integration of large language models into medical education has significantly increased, providing valuable assistance in single-turn, isolated educational tasks. However, their utility remains limited in complex, iterative instructional workflows characteristic of clinical education. Single-prompt AI chatbots lack the necessary contextual awareness and iterative capability required for nuanced educational tasks. This Viewpoint paper argues for a shift from conventional chatbot paradigms toward a modular, multistep artificial intelligence (AI) agent framework that aligns closely with the pedagogical needs of medical educators. We propose a modular framework composed of specialized AI agents, each responsible for distinct instructional subtasks. Furthermore, these agents operate within clearly defined boundaries and are equipped with tools and resources to accomplish their tasks and ensure pedagogical continuity and coherence. Specialized agents enhance accuracy by using models optimally tailored to specific cognitive tasks, increasing the quality of outputs compared to single-model workflows. Using a clinical scenario design as an illustrative example, we demonstrate how task specialization, iterative feedback, and tool integration in an agent-based pipeline can mirror expert-driven educational processes. The framework maintains a human-in-the-loop structure, with educators reviewing and refining each output before progression, ensuring pedagogical integrity, flexibility, and transparency. Our proposed shift toward modular AI agents offers significant promise for enhancing educational workflows by delegating routine tasks to specialized systems. We encourage educators to explore how these emerging AI ecosystems could transform medical education.
{"title":"Beyond Chatbots: Moving Toward Multistep Modular AI Agents in Medical Education.","authors":"Minyang Chow, Olivia Ng","doi":"10.2196/76661","DOIUrl":"10.2196/76661","url":null,"abstract":"<p><strong>Unlabelled: </strong>The integration of large language models into medical education has significantly increased, providing valuable assistance in single-turn, isolated educational tasks. However, their utility remains limited in complex, iterative instructional workflows characteristic of clinical education. Single-prompt AI chatbots lack the necessary contextual awareness and iterative capability required for nuanced educational tasks. This Viewpoint paper argues for a shift from conventional chatbot paradigms toward a modular, multistep artificial intelligence (AI) agent framework that aligns closely with the pedagogical needs of medical educators. We propose a modular framework composed of specialized AI agents, each responsible for distinct instructional subtasks. Furthermore, these agents operate within clearly defined boundaries and are equipped with tools and resources to accomplish their tasks and ensure pedagogical continuity and coherence. Specialized agents enhance accuracy by using models optimally tailored to specific cognitive tasks, increasing the quality of outputs compared to single-model workflows. Using a clinical scenario design as an illustrative example, we demonstrate how task specialization, iterative feedback, and tool integration in an agent-based pipeline can mirror expert-driven educational processes. The framework maintains a human-in-the-loop structure, with educators reviewing and refining each output before progression, ensuring pedagogical integrity, flexibility, and transparency. Our proposed shift toward modular AI agents offers significant promise for enhancing educational workflows by delegating routine tasks to specialized systems. We encourage educators to explore how these emerging AI ecosystems could transform medical education.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e76661"},"PeriodicalIF":3.2,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12490774/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145214003","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ming-Yu Hsieh, Tzu-Ling Wang, Pen-Hua Su, Ming-Chih Chou
Background: Large language models such as ChatGPT (OpenAI) have shown promise in medical education assessments, but the comparative effects of prompt engineering across optimized variants and relative performance against medical students remain unclear.
Objective: This study aims to systematically evaluate the impact of prompt engineering on five ChatGPT variants (GPT-3.5, GPT-4.0, GPT-4o, GPT-4o1-mini, and GPT-4o1) and benchmark their performance against fourth-year medical students in midterm and final examinations.
Methods: A 100-item examination dataset covering multiple choice questions, short answer questions, clinical case analysis, and image-based questions was administered to each model under no-prompt and prompt-engineering conditions over 5 independent runs. Student cohort scores (N=143) were collected for comparison. Responses were scored using standardized rubrics, converted to percentages, and analyzed in SPSS Statistics (v29.0) with paired t tests and Cohen d (P<.05).
Results: Baseline midterm scores ranged from 59.2% (GPT-3.5) to 94.1% (GPT-4o1), and final scores ranged from 55% to 92.4%. Fourth-year students averaged 89.4% (midterm) and 80.2% (final). Prompt engineering significantly improved GPT-3.5 (10.6%, P<.001) and GPT-4.0 (3.2%, P=.002) but yielded negligible gains for optimized variants (P=.07-.94). Optimized models matched or exceeded student performance on both exams.
Conclusions: Prompt engineering enhances early-generation model performance, whereas advanced variants inherently achieve near-ceiling accuracy, surpassing medical students. As large language models mature, emphasis should shift from prompt design to model selection, multimodal integration, and critical use of artificial intelligence as a learning companion.
{"title":"Impact of Prompt Engineering on the Performance of ChatGPT Variants Across Different Question Types in Medical Student Examinations: Cross-Sectional Study.","authors":"Ming-Yu Hsieh, Tzu-Ling Wang, Pen-Hua Su, Ming-Chih Chou","doi":"10.2196/78320","DOIUrl":"10.2196/78320","url":null,"abstract":"<p><strong>Background: </strong>Large language models such as ChatGPT (OpenAI) have shown promise in medical education assessments, but the comparative effects of prompt engineering across optimized variants and relative performance against medical students remain unclear.</p><p><strong>Objective: </strong>This study aims to systematically evaluate the impact of prompt engineering on five ChatGPT variants (GPT-3.5, GPT-4.0, GPT-4o, GPT-4o1-mini, and GPT-4o1) and benchmark their performance against fourth-year medical students in midterm and final examinations.</p><p><strong>Methods: </strong>A 100-item examination dataset covering multiple choice questions, short answer questions, clinical case analysis, and image-based questions was administered to each model under no-prompt and prompt-engineering conditions over 5 independent runs. Student cohort scores (N=143) were collected for comparison. Responses were scored using standardized rubrics, converted to percentages, and analyzed in SPSS Statistics (v29.0) with paired t tests and Cohen d (P<.05).</p><p><strong>Results: </strong>Baseline midterm scores ranged from 59.2% (GPT-3.5) to 94.1% (GPT-4o1), and final scores ranged from 55% to 92.4%. Fourth-year students averaged 89.4% (midterm) and 80.2% (final). Prompt engineering significantly improved GPT-3.5 (10.6%, P<.001) and GPT-4.0 (3.2%, P=.002) but yielded negligible gains for optimized variants (P=.07-.94). Optimized models matched or exceeded student performance on both exams.</p><p><strong>Conclusions: </strong>Prompt engineering enhances early-generation model performance, whereas advanced variants inherently achieve near-ceiling accuracy, surpassing medical students. As large language models mature, emphasis should shift from prompt design to model selection, multimodal integration, and critical use of artificial intelligence as a learning companion.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e78320"},"PeriodicalIF":3.2,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12488032/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145207885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background: Traditional Chinese medicine (TCM) education in China has evolved significantly, shaped by both national policy and social needs. Despite this, the academic community has yet to fully explore the long-term trends and core issues in TCM education policies. As the global interest in TCM continues to grow, understanding these trends becomes crucial for guiding future policy and educational reforms. This study used cutting-edge deep learning techniques to fill this gap, offering a novel, data-driven perspective on the evolution of TCM education policies.
Objective: This study aimed to systematically analyze the research topics and evolutionary trends in TCM education policies in China using a deep learning-based topic modeling approach, providing valuable insights to guide future policy development and educational practices.
Methods: TCM policy-related documents were collected from major sources, including the Ministry of Education, the National Administration of Traditional Chinese Medicine, PKU Lawinfo, and archives of TCM colleges. The text was preprocessed and analyzed using the BERTopic model, a state-of-the-art tool for topic modeling, to extract key themes and examine the policy development trajectory.
Results: The analysis revealed 27 core topics in TCM education policies, including medical education, curriculum reform, rural health care, internationalization, and the integration of TCM with modern education systems. These topics were clustered into 5 stages of policy evolution: marginalization, standardization, specialization, systematization, and restandardization. These stages reflect the ongoing balancing act between modernizing TCM education and preserving its traditional values, while adapting to national political, social, and economic strategies.
Conclusions: This study offers groundbreaking insights into the dynamic and multifaceted evolution of TCM education policies in China. By leveraging the BERTopic model, it provides a comprehensive framework for understanding the forces shaping TCM education and offers actionable recommendations for future policy making. The findings are essential for educators, policymakers, and researchers aiming to refine and innovate TCM education in an increasingly globalized world.
{"title":"Mapping the Evolution of China's Traditional Chinese Medicine Education Policies: Insights From a BERTopic-Based Descriptive Study.","authors":"Tao Yang, Fan Yang, Yong Li","doi":"10.2196/72660","DOIUrl":"10.2196/72660","url":null,"abstract":"<p><strong>Background: </strong>Traditional Chinese medicine (TCM) education in China has evolved significantly, shaped by both national policy and social needs. Despite this, the academic community has yet to fully explore the long-term trends and core issues in TCM education policies. As the global interest in TCM continues to grow, understanding these trends becomes crucial for guiding future policy and educational reforms. This study used cutting-edge deep learning techniques to fill this gap, offering a novel, data-driven perspective on the evolution of TCM education policies.</p><p><strong>Objective: </strong>This study aimed to systematically analyze the research topics and evolutionary trends in TCM education policies in China using a deep learning-based topic modeling approach, providing valuable insights to guide future policy development and educational practices.</p><p><strong>Methods: </strong>TCM policy-related documents were collected from major sources, including the Ministry of Education, the National Administration of Traditional Chinese Medicine, PKU Lawinfo, and archives of TCM colleges. The text was preprocessed and analyzed using the BERTopic model, a state-of-the-art tool for topic modeling, to extract key themes and examine the policy development trajectory.</p><p><strong>Results: </strong>The analysis revealed 27 core topics in TCM education policies, including medical education, curriculum reform, rural health care, internationalization, and the integration of TCM with modern education systems. These topics were clustered into 5 stages of policy evolution: marginalization, standardization, specialization, systematization, and restandardization. These stages reflect the ongoing balancing act between modernizing TCM education and preserving its traditional values, while adapting to national political, social, and economic strategies.</p><p><strong>Conclusions: </strong>This study offers groundbreaking insights into the dynamic and multifaceted evolution of TCM education policies in China. By leveraging the BERTopic model, it provides a comprehensive framework for understanding the forces shaping TCM education and offers actionable recommendations for future policy making. The findings are essential for educators, policymakers, and researchers aiming to refine and innovate TCM education in an increasingly globalized world.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e72660"},"PeriodicalIF":3.2,"publicationDate":"2025-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12463338/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145150792","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sophie Sarah Rossner, Muluken Gizaw, Sefonias Getachew, Eyerusalem Getachew, Alemnew Destaw, Sarah Negash, Lena Bauer, Eva Susanne Marion Hermann, Abel Shita, Susanne Unverzagt, Pablo Sandro Carvalho Santos, Eva Johanna Kantelhardt, Eric Sven Kroeber
Background: Training of health care professionals and their participation in continuous medical education are crucial to ensure quality health care. Low-resource countries in Sub-Saharan Africa struggle with health care disparities between urban and rural areas concerning access to educational resources. While e-learning can facilitate a wide distribution of educational content, it depends on learners' engagement and infrastructure.
Objective: This study aims to assess knowledge, attitude, practice, and access to infrastructure related to e-learning among health care professionals in primary health care settings in Ethiopia.
Methods: In April 2023, we carried out a quantitative, questionnaire-based cross-sectional study guided by the knowledge, attitudes, and practice framework, including additional items on available infrastructure. The scores in each category are defined as "high" and "low" based on the median, followed by the application of logistic regression on selected sociodemographic factors. We included health care professionals working in general and primary hospitals, health centers, and health posts.
Results: Of 398 participants (response rate 94.5%), more than half (n=207, 52%) reported feeling confident about their understanding of e-learning and conducting online searches, both for general (n=247, 62.1%) and medical-related content (n=251, 63.1%). Higher levels of education were associated with better knowledge (adjusted odds ratio [AOR] 2.32, 95% CI 1.45-3.68). Regardless of financial and personal efforts, we observed a generally positive attitude. Almost half of the participants (n=172, 43.2%) reported using the internet daily, compared to 16.8% (n=67) of participants who never used the internet. Higher education (AOR 2.56, 95% CI 1.57-4.16) and income levels (AOR 1.31, 95% CI 1.06-1.62) were associated with higher practice scores of e-learning-related activities. Women, however, exhibited lower practice scores (AOR 0.44, 95% CI 0.27-0.71). Regular access to an internet-enabled device was reported by 43.5% (n=173) of the participants. Smartphones were the primarily used device (268/393, 67.3%). Common barriers to internet access were limited internet availability (142/437, 32.5%) and costs (n=190, 43.5%). Higher education (AOR 1.56, 95% CI 0.98, 2.46) and income (AOR 1.50; 95% CI 1.21-1.85) were associated with increased access to infrastructure, while it was decreased for women (AOR 0.48, 95% CI 0.30-0.77).
Conclusions: Although Ethiopian health care professionals report mixed levels of knowledge, they have a positive attitude toward e-learning in medical education. While internet use is common, especially via smartphone, the access to devices and reliable internet is limited. To improve accessibility, investments in the digital infrastructure and individual digital education programs are necessary, especially targetin
背景:卫生保健专业人员的培训和他们参与持续医学教育是确保高质量卫生保健的关键。撒哈拉以南非洲资源匮乏国家努力解决城乡之间在获得教育资源方面的保健差距问题。虽然电子学习可以促进教育内容的广泛传播,但它取决于学习者的参与和基础设施。目的:本研究旨在评估埃塞俄比亚初级卫生保健机构中卫生保健专业人员与电子学习相关的知识、态度、做法和基础设施的获取情况。方法:在知识、态度和实践框架的指导下,我们于2023年4月进行了一项定量的、基于问卷的横断面研究,包括关于可用基础设施的附加项目。每个类别的得分根据中位数定义为“高”和“低”,然后对选定的社会人口因素应用逻辑回归。我们包括在综合医院和初级医院、卫生中心和卫生站工作的卫生保健专业人员。结果:在398名参与者(回复率94.5%)中,超过一半(n=207, 52%)表示对他们对电子学习和在线搜索的理解充满信心,包括一般内容(n=247, 62.1%)和医学相关内容(n=251, 63.1%)。较高的教育水平与更好的知识相关(调整优势比[AOR] 2.32, 95% CI 1.45-3.68)。不管经济和个人努力如何,我们观察到总体上是积极的态度。几乎一半的参与者(n=172, 43.2%)报告每天使用互联网,相比之下,16.8% (n=67)的参与者从不使用互联网。高等教育(AOR 2.56, 95% CI 1.57-4.16)和收入水平(AOR 1.31, 95% CI 1.06-1.62)与更高的电子学习相关活动实践得分相关。然而,女性表现出较低的练习得分(AOR 0.44, 95% CI 0.27-0.71)。43.5% (n=173)的参与者报告说他们经常使用能上网的设备。智能手机是主要的使用设备(268/393,67.3%)。互联网接入的常见障碍是有限的互联网可用性(142/437,32.5%)和成本(n=190, 43.5%)。高等教育(AOR为1.56,95% CI为0.98,2.46)和收入(AOR为1.50,95% CI为1.21-1.85)与获得基础设施的机会增加有关,而女性获得基础设施的机会减少(AOR为0.48,95% CI为0.30-0.77)。结论:尽管埃塞俄比亚卫生保健专业人员报告的知识水平参差不齐,但他们对医学教育中的电子学习持积极态度。虽然互联网使用很普遍,尤其是通过智能手机,但访问设备和可靠的互联网是有限的。为了改善可及性,有必要投资于数字基础设施和个人数字教育项目,特别是针对妇女和低收入群体。由于它们的广泛可用性,电子学习程序应该针对智能手机进行优化。
{"title":"Health Care Professionals' Knowledge, Attitude, Practice, and Infrastructure Accessibility for e-Learning in Ethiopia: Cross-Sectional Study.","authors":"Sophie Sarah Rossner, Muluken Gizaw, Sefonias Getachew, Eyerusalem Getachew, Alemnew Destaw, Sarah Negash, Lena Bauer, Eva Susanne Marion Hermann, Abel Shita, Susanne Unverzagt, Pablo Sandro Carvalho Santos, Eva Johanna Kantelhardt, Eric Sven Kroeber","doi":"10.2196/65598","DOIUrl":"10.2196/65598","url":null,"abstract":"<p><strong>Background: </strong>Training of health care professionals and their participation in continuous medical education are crucial to ensure quality health care. Low-resource countries in Sub-Saharan Africa struggle with health care disparities between urban and rural areas concerning access to educational resources. While e-learning can facilitate a wide distribution of educational content, it depends on learners' engagement and infrastructure.</p><p><strong>Objective: </strong>This study aims to assess knowledge, attitude, practice, and access to infrastructure related to e-learning among health care professionals in primary health care settings in Ethiopia.</p><p><strong>Methods: </strong>In April 2023, we carried out a quantitative, questionnaire-based cross-sectional study guided by the knowledge, attitudes, and practice framework, including additional items on available infrastructure. The scores in each category are defined as \"high\" and \"low\" based on the median, followed by the application of logistic regression on selected sociodemographic factors. We included health care professionals working in general and primary hospitals, health centers, and health posts.</p><p><strong>Results: </strong>Of 398 participants (response rate 94.5%), more than half (n=207, 52%) reported feeling confident about their understanding of e-learning and conducting online searches, both for general (n=247, 62.1%) and medical-related content (n=251, 63.1%). Higher levels of education were associated with better knowledge (adjusted odds ratio [AOR] 2.32, 95% CI 1.45-3.68). Regardless of financial and personal efforts, we observed a generally positive attitude. Almost half of the participants (n=172, 43.2%) reported using the internet daily, compared to 16.8% (n=67) of participants who never used the internet. Higher education (AOR 2.56, 95% CI 1.57-4.16) and income levels (AOR 1.31, 95% CI 1.06-1.62) were associated with higher practice scores of e-learning-related activities. Women, however, exhibited lower practice scores (AOR 0.44, 95% CI 0.27-0.71). Regular access to an internet-enabled device was reported by 43.5% (n=173) of the participants. Smartphones were the primarily used device (268/393, 67.3%). Common barriers to internet access were limited internet availability (142/437, 32.5%) and costs (n=190, 43.5%). Higher education (AOR 1.56, 95% CI 0.98, 2.46) and income (AOR 1.50; 95% CI 1.21-1.85) were associated with increased access to infrastructure, while it was decreased for women (AOR 0.48, 95% CI 0.30-0.77).</p><p><strong>Conclusions: </strong>Although Ethiopian health care professionals report mixed levels of knowledge, they have a positive attitude toward e-learning in medical education. While internet use is common, especially via smartphone, the access to devices and reliable internet is limited. To improve accessibility, investments in the digital infrastructure and individual digital education programs are necessary, especially targetin","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e65598"},"PeriodicalIF":3.2,"publicationDate":"2025-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12463343/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145150821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
<p><strong>Background: </strong>Burnout among emergency room health care workers (HCWs) has reached critical levels, affecting up to 43% of HCWs and 35% of emergency medicine personnel during the COVID-19 pandemic. Nurses were most affected, followed by physicians, leading to absenteeism, reduced care quality, and turnover rates as high as 78% in some settings such as Thailand. Beyond workforce instability, burnout compromises patient safety. Each 1-unit increase in emotional exhaustion has been linked to a 2.63-fold rise in reports of poor care quality, 30% increase in patient falls, 47% increase in medication errors, and 32% increase in health care-associated infections. Burnout is also associated with lower job satisfaction, worsening mental health, and increased intent to leave the profession. These findings underscore the urgent need for effective strategies to reduce stress and burnout in emergency care.</p><p><strong>Objective: </strong>This study aimed to evaluate the effectiveness and effect size of a multimodal learning approach-Emergency Room Virtual Simulation Interprofessional Education (ER-VIPE)-that integrates medical movies, massive online open courses (MOOCs), and computer- or virtual reality (VR)-based simulations with co-debriefing for reducing burnout and stress among future health care professionals compared with approaches lacking co-debriefing or using only movies and MOOCs.</p><p><strong>Methods: </strong>A single-blind, quasi-experimental study was conducted at a university hospital from August 2022 to September 2023 using a 3-group treatment design. Group A (control) participated in a 3D computer-based, simulation-based interprofessional education (SIMBIE) without debriefing. Group B received the ER-VIPE intervention. Group C received the same as Group B, but the computer-based SIMBIE was replaced with 3D VR-SIMBIE. SIMBIE activities simulated a COVID-19 pneumonia crisis. Outcomes included the Dundee Stress State Questionnaire (DSSQ) and the Copenhagen Burnout Inventory, with trait anxiety as a behavioral control. Stress and burnout were measured at baseline, pre-intervention, postintervention, and 1-month follow-up. Generalized estimating equations were used to analyze group differences, with statistical significance set at P<.05.</p><p><strong>Results: </strong>We randomized 87 undergraduate students from various health programs into the 3 groups (n=29 each). Participants' mean age was 22 years, with 71% (62/87) as women. After the 1-month post-SIMBIE follow-up, adjusted analyses revealed positive trends in DSSQ-engagement across all groups, with Group B showing a significant increase compared with Group A (mean difference=3.93; P=.001). DSSQ-worry and DSSQ-distress scores decreased nonsignificantly across all groups. Burnout scores also improved across groups, with Group B having a significantly lower score than Group A (mean difference=-2.02; P=.02). No significant burnout differences were found between Group C and G
{"title":"Comparing the Effectiveness of Multimodal Learning Using Computer-Based and Immersive Virtual Reality Simulation-Based Interprofessional Education With Co-Debriefing, Medical Movies, and Massive Online Open Courses for Mitigating Stress and Long-Term Burnout in Medical Training: Quasi-Experimental Study.","authors":"Sirikanyawan Srikasem, Sunisa Seephom, Atthaphon Viriyopase, Phanupong Phutrakool, Sirhavich Khowinthaseth, Khuansiri Narajeenron","doi":"10.2196/70726","DOIUrl":"10.2196/70726","url":null,"abstract":"<p><strong>Background: </strong>Burnout among emergency room health care workers (HCWs) has reached critical levels, affecting up to 43% of HCWs and 35% of emergency medicine personnel during the COVID-19 pandemic. Nurses were most affected, followed by physicians, leading to absenteeism, reduced care quality, and turnover rates as high as 78% in some settings such as Thailand. Beyond workforce instability, burnout compromises patient safety. Each 1-unit increase in emotional exhaustion has been linked to a 2.63-fold rise in reports of poor care quality, 30% increase in patient falls, 47% increase in medication errors, and 32% increase in health care-associated infections. Burnout is also associated with lower job satisfaction, worsening mental health, and increased intent to leave the profession. These findings underscore the urgent need for effective strategies to reduce stress and burnout in emergency care.</p><p><strong>Objective: </strong>This study aimed to evaluate the effectiveness and effect size of a multimodal learning approach-Emergency Room Virtual Simulation Interprofessional Education (ER-VIPE)-that integrates medical movies, massive online open courses (MOOCs), and computer- or virtual reality (VR)-based simulations with co-debriefing for reducing burnout and stress among future health care professionals compared with approaches lacking co-debriefing or using only movies and MOOCs.</p><p><strong>Methods: </strong>A single-blind, quasi-experimental study was conducted at a university hospital from August 2022 to September 2023 using a 3-group treatment design. Group A (control) participated in a 3D computer-based, simulation-based interprofessional education (SIMBIE) without debriefing. Group B received the ER-VIPE intervention. Group C received the same as Group B, but the computer-based SIMBIE was replaced with 3D VR-SIMBIE. SIMBIE activities simulated a COVID-19 pneumonia crisis. Outcomes included the Dundee Stress State Questionnaire (DSSQ) and the Copenhagen Burnout Inventory, with trait anxiety as a behavioral control. Stress and burnout were measured at baseline, pre-intervention, postintervention, and 1-month follow-up. Generalized estimating equations were used to analyze group differences, with statistical significance set at P<.05.</p><p><strong>Results: </strong>We randomized 87 undergraduate students from various health programs into the 3 groups (n=29 each). Participants' mean age was 22 years, with 71% (62/87) as women. After the 1-month post-SIMBIE follow-up, adjusted analyses revealed positive trends in DSSQ-engagement across all groups, with Group B showing a significant increase compared with Group A (mean difference=3.93; P=.001). DSSQ-worry and DSSQ-distress scores decreased nonsignificantly across all groups. Burnout scores also improved across groups, with Group B having a significantly lower score than Group A (mean difference=-2.02; P=.02). No significant burnout differences were found between Group C and G","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e70726"},"PeriodicalIF":3.2,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12508677/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145139001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Unlabelled: This paper aims to describe the cocreation and development processes of an educational ecosystem-centered Bachelor's degree in Digital Health and Biomedical Innovation (SauD InoB). This program is shaped by a multidisciplinary, intersectoral, and collaborative framework, involving more than 60 organizations in teaching activities, internship supervision, or hosting, most of which collaborated in needs assessment, curriculum development, and public promotion of the degree. In the context of health care digital transformation, this comprehensive Bachelor's degree will respond to unmet demands of the labor market by training students with technological, research, and management skills, as well as with basic clinical and biomedical concepts. Graduates will become transdisciplinary, creative professionals capable of understanding and integrating different "languages," reasoning, clinical processes, and scenarios.
{"title":"An Ecosystem Approach to Developing and Implementing a Cocreated Bachelor's Degree in Digital Health and Biomedical Innovation.","authors":"Patrícia Alves, Elisio Costa, Altamiro Costa-Pereira, Inês Falcão-Pires, João Fonseca, Adelino Leite-Moreira, Bernardo Sousa-Pinto, Nuno Vale","doi":"10.2196/63903","DOIUrl":"10.2196/63903","url":null,"abstract":"<p><strong>Unlabelled: </strong>This paper aims to describe the cocreation and development processes of an educational ecosystem-centered Bachelor's degree in Digital Health and Biomedical Innovation (SauD InoB). This program is shaped by a multidisciplinary, intersectoral, and collaborative framework, involving more than 60 organizations in teaching activities, internship supervision, or hosting, most of which collaborated in needs assessment, curriculum development, and public promotion of the degree. In the context of health care digital transformation, this comprehensive Bachelor's degree will respond to unmet demands of the labor market by training students with technological, research, and management skills, as well as with basic clinical and biomedical concepts. Graduates will become transdisciplinary, creative professionals capable of understanding and integrating different \"languages,\" reasoning, clinical processes, and scenarios.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e63903"},"PeriodicalIF":3.2,"publicationDate":"2025-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12457801/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145132010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background: The integration of digital technology in dental education has been recognized for its potential to address the challenges in training removable partial denture (RPD) design. RPD framework design is crucial to long-term success in the treatment of dentition defects, but traditional training methods often fall short of adequately preparing students for real-world applications.
Objective: This study aimed to evaluate the efficacy of a 3D simulation-based preclinical training software for RPDs in enhancing learning outcomes among first-year stomatology master's students, while also assessing user perceptions among students and faculty.
Methods: RTS (Yikchi Siu) is a preclinical training software that simulates the clinical process of treating patients with partial edentulism. In this study, 26 newly enrolled master's degree students in stomatology who volunteered to participate were randomly divided into a control group (n=13) and a training group (n=13). The training group used the RTS for 2 credit hours (90 min) of self-study, while the control group received theoretical lessons and case practice from an instructor. After 2 hours, both groups completed the theoretical knowledge and drawing tests for RPD simultaneously. Test results were evaluated and graded by 2 experts in prosthodontics. Both users and teachers filled out a questionnaire afterward about their training experience.
Results: Participants in the training group obtained better final grades compared to controls (theoretical test: 88.8, SD 2.3; 85.7, SD 3.3, respectively; P=.01; drawing test: 89.8, SD 4.5; 85.1, SD 4.3, respectively; P=.01). The training group had a shorter completion time in the drawing test (12.6, SD 19 min; 17.7, SD 3 min, respectively; P<.001) but there were no significant differences in the completion times in the theoretical test (23.2, SD 2.2 min; 24.9, SD 2.8 min, respectively; P=.14). Students and faculty generally had a favorable opinion of the RTS.
Conclusions: The effectiveness of the RTS for newly enrolled master's degree students in stomatology to understand and apply their knowledge of RPD framework design was validated; the system was well received by both students and faculty members, who reported that it improved the effectiveness and convenience of teaching.
{"title":"Enhancing Preclinical Training for Removable Partial Dentures Through Participatory 3D Simulation: Development and Usability Study.","authors":"Yikchi Siu, Hefei Bai, Jung-Min Yoon, Hongqiang Ye, Yunsong Liu, Yongsheng Zhou","doi":"10.2196/71743","DOIUrl":"10.2196/71743","url":null,"abstract":"<p><strong>Background: </strong>The integration of digital technology in dental education has been recognized for its potential to address the challenges in training removable partial denture (RPD) design. RPD framework design is crucial to long-term success in the treatment of dentition defects, but traditional training methods often fall short of adequately preparing students for real-world applications.</p><p><strong>Objective: </strong>This study aimed to evaluate the efficacy of a 3D simulation-based preclinical training software for RPDs in enhancing learning outcomes among first-year stomatology master's students, while also assessing user perceptions among students and faculty.</p><p><strong>Methods: </strong>RTS (Yikchi Siu) is a preclinical training software that simulates the clinical process of treating patients with partial edentulism. In this study, 26 newly enrolled master's degree students in stomatology who volunteered to participate were randomly divided into a control group (n=13) and a training group (n=13). The training group used the RTS for 2 credit hours (90 min) of self-study, while the control group received theoretical lessons and case practice from an instructor. After 2 hours, both groups completed the theoretical knowledge and drawing tests for RPD simultaneously. Test results were evaluated and graded by 2 experts in prosthodontics. Both users and teachers filled out a questionnaire afterward about their training experience.</p><p><strong>Results: </strong>Participants in the training group obtained better final grades compared to controls (theoretical test: 88.8, SD 2.3; 85.7, SD 3.3, respectively; P=.01; drawing test: 89.8, SD 4.5; 85.1, SD 4.3, respectively; P=.01). The training group had a shorter completion time in the drawing test (12.6, SD 19 min; 17.7, SD 3 min, respectively; P<.001) but there were no significant differences in the completion times in the theoretical test (23.2, SD 2.2 min; 24.9, SD 2.8 min, respectively; P=.14). Students and faculty generally had a favorable opinion of the RTS.</p><p><strong>Conclusions: </strong>The effectiveness of the RTS for newly enrolled master's degree students in stomatology to understand and apply their knowledge of RPD framework design was validated; the system was well received by both students and faculty members, who reported that it improved the effectiveness and convenience of teaching.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e71743"},"PeriodicalIF":3.2,"publicationDate":"2025-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12454677/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145125998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Anila Jaleel, Umair Aziz, Ghulam Farid, Muhammad Zahid Bashir, Tehmasp Rehman Mirza, Syed Mohammad Khizar Abbas, Shiraz Aslam, Rana Muhammad Hassaan Sikander
Background: Artificial intelligence (AI) has significantly impacted health care, medicine, and radiology, offering personalized treatment plans, simplified workflows, and informed clinical decisions. ChatGPT (OpenAI), a conversational AI model, has revolutionized health care and medical education by simulating clinical scenarios and improving communication skills. However, inconsistent performance across medical licensing examinations and variability between countries and specialties highlight the need for further research on contextual factors influencing AI accuracy and exploring its potential to enhance technical proficiency and soft skills, making AI a reliable tool in patient care and medical education.
Objective: This systematic review aims to evaluate and compare the accuracy and potential of ChatGPT-3.5 and 4.0 in medical licensing and in-training residency examinations across various countries and specialties.
Methods: A systematic review and meta-analysis were conducted, adhering to the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines. Data were collected from multiple reputable databases (Scopus, PubMed, JMIR Publications, Elsevier, BMJ, and Wiley Online Library), focusing on studies published from January 2023 to July 2024. Analysis specifically targeted research assessing ChatGPT's efficacy in medical licensing exams, excluding studies not related to this focus or published in languages other than English. Ultimately, 53 studies were included, providing a robust dataset for comparing the accuracy rates of ChatGPT-3.5 and 4.0.
Results: ChatGPT-4 outperformed ChatGPT-3.5 in medical licensing exams, achieving a pooled accuracy of 81.8%, compared to ChatGPT-3.5's 60.8%. In in-training residency exams, ChatGPT-4 achieved an accuracy rate of 72.2%, compared to 57.7% for ChatGPT-3.5. The forest plot presented a risk ratio of 1.36 (95% CI 1.30-1.43), demonstrating that ChatGPT-4 was 36% more likely to provide correct answers than ChatGPT-3.5 across both medical licensing and residency exams. These results indicate that ChatGPT-4 significantly outperforms ChatGPT-3.5, but the performance advantage varies depending on the exam type. This highlights the importance of targeted improvements and further research to optimize ChatGPT-4's performance in specific educational and clinical settings.
Conclusions: ChatGPT-4.0 and 3.5 show promising results in enhancing medical education and supporting clinical decision-making, but they cannot replace the comprehensive skill set required for effective medical practice. Future research should focus on improving AI's capabilities in interpreting complex clinical data and enhancing its reliability as an educational resource.
背景:人工智能(AI)对医疗保健、医学和放射学产生了重大影响,提供了个性化的治疗计划、简化的工作流程和知情的临床决策。ChatGPT (OpenAI)是一种会话式人工智能模型,通过模拟临床场景和提高沟通技巧,彻底改变了医疗保健和医学教育。然而,在医疗执照考试中的不一致表现以及国家和专业之间的差异突出表明,需要进一步研究影响人工智能准确性的背景因素,并探索其提高技术熟练程度和软技能的潜力,使人工智能成为患者护理和医学教育中的可靠工具。目的:本系统综述旨在评估和比较ChatGPT-3.5和4.0在不同国家和专业的医疗许可和培训住院医师考试中的准确性和潜力。方法:遵循PRISMA(首选系统评价和荟萃分析报告项目)指南进行系统评价和荟萃分析。数据收集自多个知名数据库(Scopus、PubMed、JMIR Publications、Elsevier、BMJ和Wiley Online Library),重点关注2023年1月至2024年7月发表的研究。分析专门针对评估ChatGPT在医疗执照考试中的疗效的研究,排除与该重点无关或以英语以外的语言发表的研究。最终,纳入了53项研究,为比较ChatGPT-3.5和4.0的准确率提供了一个强大的数据集。结果:ChatGPT-4在医疗执照考试中的表现优于ChatGPT-3.5,达到81.8%的汇总准确率,而ChatGPT-3.5的准确率为60.8%。在实习医师考试中,ChatGPT-4的准确率为72.2%,而ChatGPT-3.5的准确率为57.7%。森林图的风险比为1.36 (95% CI 1.30-1.43),表明在医疗执照和住院医师考试中,ChatGPT-4提供正确答案的可能性比ChatGPT-3.5高36%。这些结果表明,ChatGPT-4明显优于ChatGPT-3.5,但性能优势因考试类型而异。这突出了有针对性的改进和进一步研究的重要性,以优化ChatGPT-4在特定教育和临床环境中的表现。结论:ChatGPT-4.0和3.5在加强医学教育和支持临床决策方面显示出有希望的结果,但它们不能取代有效医疗实践所需的综合技能。未来的研究应侧重于提高人工智能在解释复杂临床数据方面的能力,并提高其作为教育资源的可靠性。
{"title":"Evaluating the Potential and Accuracy of ChatGPT-3.5 and 4.0 in Medical Licensing and In-Training Examinations: Systematic Review and Meta-Analysis.","authors":"Anila Jaleel, Umair Aziz, Ghulam Farid, Muhammad Zahid Bashir, Tehmasp Rehman Mirza, Syed Mohammad Khizar Abbas, Shiraz Aslam, Rana Muhammad Hassaan Sikander","doi":"10.2196/68070","DOIUrl":"10.2196/68070","url":null,"abstract":"<p><strong>Background: </strong>Artificial intelligence (AI) has significantly impacted health care, medicine, and radiology, offering personalized treatment plans, simplified workflows, and informed clinical decisions. ChatGPT (OpenAI), a conversational AI model, has revolutionized health care and medical education by simulating clinical scenarios and improving communication skills. However, inconsistent performance across medical licensing examinations and variability between countries and specialties highlight the need for further research on contextual factors influencing AI accuracy and exploring its potential to enhance technical proficiency and soft skills, making AI a reliable tool in patient care and medical education.</p><p><strong>Objective: </strong>This systematic review aims to evaluate and compare the accuracy and potential of ChatGPT-3.5 and 4.0 in medical licensing and in-training residency examinations across various countries and specialties.</p><p><strong>Methods: </strong>A systematic review and meta-analysis were conducted, adhering to the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines. Data were collected from multiple reputable databases (Scopus, PubMed, JMIR Publications, Elsevier, BMJ, and Wiley Online Library), focusing on studies published from January 2023 to July 2024. Analysis specifically targeted research assessing ChatGPT's efficacy in medical licensing exams, excluding studies not related to this focus or published in languages other than English. Ultimately, 53 studies were included, providing a robust dataset for comparing the accuracy rates of ChatGPT-3.5 and 4.0.</p><p><strong>Results: </strong>ChatGPT-4 outperformed ChatGPT-3.5 in medical licensing exams, achieving a pooled accuracy of 81.8%, compared to ChatGPT-3.5's 60.8%. In in-training residency exams, ChatGPT-4 achieved an accuracy rate of 72.2%, compared to 57.7% for ChatGPT-3.5. The forest plot presented a risk ratio of 1.36 (95% CI 1.30-1.43), demonstrating that ChatGPT-4 was 36% more likely to provide correct answers than ChatGPT-3.5 across both medical licensing and residency exams. These results indicate that ChatGPT-4 significantly outperforms ChatGPT-3.5, but the performance advantage varies depending on the exam type. This highlights the importance of targeted improvements and further research to optimize ChatGPT-4's performance in specific educational and clinical settings.</p><p><strong>Conclusions: </strong>ChatGPT-4.0 and 3.5 show promising results in enhancing medical education and supporting clinical decision-making, but they cannot replace the comprehensive skill set required for effective medical practice. Future research should focus on improving AI's capabilities in interpreting complex clinical data and enhancing its reliability as an educational resource.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e68070"},"PeriodicalIF":3.2,"publicationDate":"2025-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12495368/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145092539","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background: Artificial intelligence (AI) has revolutionized medical education by delivering tools that enhance and optimize learning. However, there is limited research on the medical students' perceptions regarding the effectiveness of AI as a learning tool, particularly in Sri Lanka.
Objective: The study aimed to explore students' perceived barriers and limitations to using AI for learning as well as their expectations in terms of future use of AI in medical education.
Methods: An exploratory qualitative study was conducted in September 2024, involving focus group discussions with medical students from two major universities in Sri Lanka. Reflexive thematic analysis was used to identify key themes and subthemes emerging from the discussions.
Results: Thirty-eight medical students participated in 5 focus group discussions. The majority of the participants were Sinhalese female students. The perceived benefits included saving time and effort and collecting and summarizing information. However, concerns and limitations centered around inaccuracies of information provided and the negative impacts on critical thinking, social interactions (peer and student teacher), and long-term retention of knowledge. Students were confused about contradictory messages received from educators regarding the use of AI for teaching and learning. However, participants showed an enthusiasm for learning more about the ethical use of AI to enhance learning and indicated that basic AI knowledge should be taught in their undergraduate program.
Conclusions: Participants recognized several benefits of AI-assisted learning but also expressed concerns and limitations requiring further studies for effective integration of AI into medical education. They expressed openness and enthusiasm for using AI while demonstrating confusion and reluctance due to the perspectives and stance of educators. We recommend educating both the educators and learners on the ethical use of AI, enabling a formal integration of AI tools into medical curricula.
{"title":"Perception of Medical Undergraduates on Artificial Intelligence in Medical Education: Qualitative Exploration.","authors":"Thilanka Seneviratne, Kaumudee Kodikara, Isuru Abeykoon, Wathsala Palpola","doi":"10.2196/73798","DOIUrl":"10.2196/73798","url":null,"abstract":"<p><strong>Background: </strong>Artificial intelligence (AI) has revolutionized medical education by delivering tools that enhance and optimize learning. However, there is limited research on the medical students' perceptions regarding the effectiveness of AI as a learning tool, particularly in Sri Lanka.</p><p><strong>Objective: </strong>The study aimed to explore students' perceived barriers and limitations to using AI for learning as well as their expectations in terms of future use of AI in medical education.</p><p><strong>Methods: </strong>An exploratory qualitative study was conducted in September 2024, involving focus group discussions with medical students from two major universities in Sri Lanka. Reflexive thematic analysis was used to identify key themes and subthemes emerging from the discussions.</p><p><strong>Results: </strong>Thirty-eight medical students participated in 5 focus group discussions. The majority of the participants were Sinhalese female students. The perceived benefits included saving time and effort and collecting and summarizing information. However, concerns and limitations centered around inaccuracies of information provided and the negative impacts on critical thinking, social interactions (peer and student teacher), and long-term retention of knowledge. Students were confused about contradictory messages received from educators regarding the use of AI for teaching and learning. However, participants showed an enthusiasm for learning more about the ethical use of AI to enhance learning and indicated that basic AI knowledge should be taught in their undergraduate program.</p><p><strong>Conclusions: </strong>Participants recognized several benefits of AI-assisted learning but also expressed concerns and limitations requiring further studies for effective integration of AI into medical education. They expressed openness and enthusiasm for using AI while demonstrating confusion and reluctance due to the perspectives and stance of educators. We recommend educating both the educators and learners on the ethical use of AI, enabling a formal integration of AI tools into medical curricula.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"11 ","pages":"e73798"},"PeriodicalIF":3.2,"publicationDate":"2025-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12448566/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145092535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}