Amina Ahmad, Janisha Kavumpurath, Raheesa Kader, Muna Altamimi, Monia El Hajj, Fatma Refaat Ahmed Ahmed, Muhammad Arsyad Subu, Taliaa Yafei, Hind Rashed Ali, Idil Saleh, Nabeel Al-Yateem
Unlabelled: Limited clinical placements for mental health courses in the United Arab Emirates have made it difficult to provide consistent experiential learning for undergraduate nursing students. As a result, nurse educators are considering technology-enabled learning approaches to deliver clinical skills training. This Viewpoint presents a reflective, theory-informed account of the first-year integration of an artificial intelligence (AI)-enabled, voice-interactive simulated patient into an undergraduate mental health nursing practicum. Grounded in Kolb's experiential learning cycle and aligned with established simulation best practices, the initiative was designed to support therapeutic communication, psychiatric assessment, and clinical reasoning through structured prebriefing, immersive interaction, and guided debriefing. The paper describes the educational rationale, scenario development, implementation processes, and contextual challenges encountered during real-world deployment across university and clinical environments. AI-supported simulations offered a standardized and psychologically safe context for students to engage with complex psychiatric scenarios, particularly when direct patient interaction is constrained. We discuss operational insights related to technical reliability, environmental requirements, faculty preparation, and assessment integration alongside considerations for scalability and sustainability in resource-limited settings. While AI-supported objective structured clinical examinations have been incorporated to support assessment consistency, formal psychometric validation and outcome comparisons have not been undertaken at this stage. By sharing lessons learned from early implementation, this Viewpoint contributes practical insights for nursing educators facing similar structural constraints. AI-enabled simulation is presented as a strategic complement to, rather than a replacement for, traditional clinical placements, with future empirical research needed to evaluate educational outcomes and long-term impact.
{"title":"Voices of Innovation: Reflective Report on Integrating Artificial Intelligence-Simulated Mental Health Patient Scenarios Into Undergraduate Nursing Education in the United Arab Emirates.","authors":"Amina Ahmad, Janisha Kavumpurath, Raheesa Kader, Muna Altamimi, Monia El Hajj, Fatma Refaat Ahmed Ahmed, Muhammad Arsyad Subu, Taliaa Yafei, Hind Rashed Ali, Idil Saleh, Nabeel Al-Yateem","doi":"10.2196/78161","DOIUrl":"10.2196/78161","url":null,"abstract":"<p><strong>Unlabelled: </strong>Limited clinical placements for mental health courses in the United Arab Emirates have made it difficult to provide consistent experiential learning for undergraduate nursing students. As a result, nurse educators are considering technology-enabled learning approaches to deliver clinical skills training. This Viewpoint presents a reflective, theory-informed account of the first-year integration of an artificial intelligence (AI)-enabled, voice-interactive simulated patient into an undergraduate mental health nursing practicum. Grounded in Kolb's experiential learning cycle and aligned with established simulation best practices, the initiative was designed to support therapeutic communication, psychiatric assessment, and clinical reasoning through structured prebriefing, immersive interaction, and guided debriefing. The paper describes the educational rationale, scenario development, implementation processes, and contextual challenges encountered during real-world deployment across university and clinical environments. AI-supported simulations offered a standardized and psychologically safe context for students to engage with complex psychiatric scenarios, particularly when direct patient interaction is constrained. We discuss operational insights related to technical reliability, environmental requirements, faculty preparation, and assessment integration alongside considerations for scalability and sustainability in resource-limited settings. While AI-supported objective structured clinical examinations have been incorporated to support assessment consistency, formal psychometric validation and outcome comparisons have not been undertaken at this stage. By sharing lessons learned from early implementation, this Viewpoint contributes practical insights for nursing educators facing similar structural constraints. AI-enabled simulation is presented as a strategic complement to, rather than a replacement for, traditional clinical placements, with future empirical research needed to evaluate educational outcomes and long-term impact.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e78161"},"PeriodicalIF":3.2,"publicationDate":"2026-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12981539/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147445444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Simone Mingels, Hannah Piehl, Madeline Therrien, Ekaterina Akhmad, Anniek van Hienen, Johan van Soest, Laura Hochstenbach, Andre Dekker, Olga Damman, Rianne Fijten
<p><strong>Background: </strong>Advancements in artificial intelligence (AI) are transforming health care, particularly through AI-driven clinical decision support systems (AI-CDSS) that aid in predicting disease progression and personalizing treatment. Despite their potential, adoption remains limited due to clinician concerns about algorithm misuse, misinterpretation, and lack of transparency.</p><p><strong>Objective: </strong>This qualitative study explores the informational needs and preferences of clinicians to better understand and appropriately use AI-CDSS in decision-making. In parallel, this study explores AI experts' perspectives on what information should be communicated to enable safe and appropriate use of AI-CDSS.</p><p><strong>Methods: </strong>A qualitative description design study was conducted using semistructured interviews with 16 participants (8 clinicians and 8 AI experts). Discussions focused on experiences with AI, informational needs, and feedback on existing reporting standards, including Model Cards, Model Facts, and the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis-Artificial Intelligence (TRIPOD-AI) checklist. The transcripts were analyzed through codebook thematic analysis.</p><p><strong>Results: </strong>Four key themes were identified: (1) clinicians need clear information on training data, its origin, size, and inclusion and exclusion criteria, to judge model applicability; (2) performance metrics must go beyond the area under the curve (AUC) and be clinically relevant to support informed decisions; (3) limitations and warnings about inappropriate use should be specific and clearly communicated to prevent misuse; and (4) information should be presented in layered, customizable formats within existing clinical software, avoiding unnecessary jargon, and allowing optional deeper explanations. While each of the reviewed reporting standards offered strengths, none were considered sufficient alone. Participants recommended a combined and clinician-centered approach to information delivery. Alignment of reporting standards with clinical workflows and decision thresholds was thought to be crucial to bridge the usability gap.</p><p><strong>Conclusions: </strong>To improve AI-CDSS adoption in clinical practice, reporting standards must be designed for better clinician comprehension and usability. Enhancing transparency, particularly regarding training data and performance, can likely help clinicians assess AI-CDSS more effectively. Information should be delivered in an accessible, layered format, fitting clinical workflows. Co-creation with clinicians throughout AI-CDSS development was a cross-cutting theme, highlighting its importance in ensuring tools are not only technically sound but also practically usable. Future research should explore how to structurally report on performance and validation metrics for clinician understanding and assess the impact of information prov
{"title":"Understanding Clinicians' Informational Needs for AI-Driven Clinical Decision Support Systems: Qualitative Interview Study.","authors":"Simone Mingels, Hannah Piehl, Madeline Therrien, Ekaterina Akhmad, Anniek van Hienen, Johan van Soest, Laura Hochstenbach, Andre Dekker, Olga Damman, Rianne Fijten","doi":"10.2196/85228","DOIUrl":"10.2196/85228","url":null,"abstract":"<p><strong>Background: </strong>Advancements in artificial intelligence (AI) are transforming health care, particularly through AI-driven clinical decision support systems (AI-CDSS) that aid in predicting disease progression and personalizing treatment. Despite their potential, adoption remains limited due to clinician concerns about algorithm misuse, misinterpretation, and lack of transparency.</p><p><strong>Objective: </strong>This qualitative study explores the informational needs and preferences of clinicians to better understand and appropriately use AI-CDSS in decision-making. In parallel, this study explores AI experts' perspectives on what information should be communicated to enable safe and appropriate use of AI-CDSS.</p><p><strong>Methods: </strong>A qualitative description design study was conducted using semistructured interviews with 16 participants (8 clinicians and 8 AI experts). Discussions focused on experiences with AI, informational needs, and feedback on existing reporting standards, including Model Cards, Model Facts, and the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis-Artificial Intelligence (TRIPOD-AI) checklist. The transcripts were analyzed through codebook thematic analysis.</p><p><strong>Results: </strong>Four key themes were identified: (1) clinicians need clear information on training data, its origin, size, and inclusion and exclusion criteria, to judge model applicability; (2) performance metrics must go beyond the area under the curve (AUC) and be clinically relevant to support informed decisions; (3) limitations and warnings about inappropriate use should be specific and clearly communicated to prevent misuse; and (4) information should be presented in layered, customizable formats within existing clinical software, avoiding unnecessary jargon, and allowing optional deeper explanations. While each of the reviewed reporting standards offered strengths, none were considered sufficient alone. Participants recommended a combined and clinician-centered approach to information delivery. Alignment of reporting standards with clinical workflows and decision thresholds was thought to be crucial to bridge the usability gap.</p><p><strong>Conclusions: </strong>To improve AI-CDSS adoption in clinical practice, reporting standards must be designed for better clinician comprehension and usability. Enhancing transparency, particularly regarding training data and performance, can likely help clinicians assess AI-CDSS more effectively. Information should be delivered in an accessible, layered format, fitting clinical workflows. Co-creation with clinicians throughout AI-CDSS development was a cross-cutting theme, highlighting its importance in ensuring tools are not only technically sound but also practically usable. Future research should explore how to structurally report on performance and validation metrics for clinician understanding and assess the impact of information prov","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e85228"},"PeriodicalIF":3.2,"publicationDate":"2026-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12989292/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147463727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Darci L Lammers, Jeffrey B Geske, Jane A Linderbaum, Michael W Cullen
This mixed methods pilot study evaluates the feasibility and effectiveness of microlearning for faculty development in cardiovascular education. Microlearning appears feasible and well-received for faculty development, offering a scalable, flexible approach.
{"title":"Evaluating Microlearning for Faculty Development in Medical Education: Mixed Methods Pilot Study.","authors":"Darci L Lammers, Jeffrey B Geske, Jane A Linderbaum, Michael W Cullen","doi":"10.2196/87980","DOIUrl":"10.2196/87980","url":null,"abstract":"<p><p>This mixed methods pilot study evaluates the feasibility and effectiveness of microlearning for faculty development in cardiovascular education. Microlearning appears feasible and well-received for faculty development, offering a scalable, flexible approach.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":" ","pages":"e87980"},"PeriodicalIF":3.2,"publicationDate":"2026-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146228960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sumana Mn, Supreeta R Shettar, Yogeesh D Maheshwarappa, G K Megha, Veerabhadra Swamy Gs, Chinchana Shylaja Eshwar, Shruthi Shree Sc
Background: Antimicrobial resistance (AMR) is a global problem. Training health care professionals in the rational use of antimicrobials is essential to curb AMR.
Objective: To support efforts to reduce antibiotic resistance, this study assesses how well a gamified educational intervention might improve health care professionals' and students' understanding and use of appropriate antibiotics.
Methods: This is a prospective interventional study conducted for clinical practitioners, undergraduates (MBBS and interns), postgraduates, and pharmacy students. A total of 60 participants were included in the study. Innovative games were administered to support the management of infections across multiple body systems, in accordance with the 2022 Indian Council of Medical Research treatment guidelines and the latest Infectious Diseases Society of America guidelines, incorporating multiple instructional components. Pretest and posttest questionnaires were administered and evaluated.
Results: After the intervention, participants' ability to differentiate between bacterial and viral symptoms in respiratory tract infections and gastroenteritis improved from 48% to 94%. The practice of selecting the appropriate empirical antimicrobial at the correct dose, route, and duration also demonstrated significant improvement, reflecting enhanced adherence to principles of rational antimicrobial use.
Conclusions: The gamified intervention successfully improved participants' knowledge and awareness of rational antimicrobial use. Substantial improvements across all the assessed components highlight the positive impact of the intervention in promoting optimal antimicrobial use and curbing AMR. Innovative gamified interventions may foster better and longer-lasting awareness, supporting appropriate antimicrobial use.
{"title":"Targeted Educational Intervention Through Game-Based Learning to Promote Rational Antimicrobial Use Among Health Care Learners: Prospective Interventional Study.","authors":"Sumana Mn, Supreeta R Shettar, Yogeesh D Maheshwarappa, G K Megha, Veerabhadra Swamy Gs, Chinchana Shylaja Eshwar, Shruthi Shree Sc","doi":"10.2196/72236","DOIUrl":"https://doi.org/10.2196/72236","url":null,"abstract":"<p><strong>Background: </strong>Antimicrobial resistance (AMR) is a global problem. Training health care professionals in the rational use of antimicrobials is essential to curb AMR.</p><p><strong>Objective: </strong>To support efforts to reduce antibiotic resistance, this study assesses how well a gamified educational intervention might improve health care professionals' and students' understanding and use of appropriate antibiotics.</p><p><strong>Methods: </strong>This is a prospective interventional study conducted for clinical practitioners, undergraduates (MBBS and interns), postgraduates, and pharmacy students. A total of 60 participants were included in the study. Innovative games were administered to support the management of infections across multiple body systems, in accordance with the 2022 Indian Council of Medical Research treatment guidelines and the latest Infectious Diseases Society of America guidelines, incorporating multiple instructional components. Pretest and posttest questionnaires were administered and evaluated.</p><p><strong>Results: </strong>After the intervention, participants' ability to differentiate between bacterial and viral symptoms in respiratory tract infections and gastroenteritis improved from 48% to 94%. The practice of selecting the appropriate empirical antimicrobial at the correct dose, route, and duration also demonstrated significant improvement, reflecting enhanced adherence to principles of rational antimicrobial use.</p><p><strong>Conclusions: </strong>The gamified intervention successfully improved participants' knowledge and awareness of rational antimicrobial use. Substantial improvements across all the assessed components highlight the positive impact of the intervention in promoting optimal antimicrobial use and curbing AMR. Innovative gamified interventions may foster better and longer-lasting awareness, supporting appropriate antimicrobial use.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e72236"},"PeriodicalIF":3.2,"publicationDate":"2026-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12974994/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147436326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background: Although artificial intelligence (AI) is being rapidly integrated into medical education, insights from medical students, particularly in the Chinese context, remain limited.
Objective: This study was designed to explore Chinese medical students' perceptions of and attitudes toward the integration of AI into medical education, as well as the factors that may influence their perspectives. The findings of this research offer valuable insights to assist medical educators in the effective implementation of these innovative educational approaches.
Methods: On the basis of the estimated total number of clinical medical students at the target institutions, the sample size was calculated to be 379. A web-based questionnaire survey was designed to investigate the acceptance level of medical students toward the application of AI. The questionnaire consisted of 14 questions across 4 dimensions, which included demographic characteristics, perceptions of AI application, willingness, and concerns. Each dimension contained 3 to 4 questions. Descriptive statistics were used to tabulate the frequency of each variable. Chi-square tests and multiple regression analyses were conducted to measure the influencing factors.
Results: A total of 566 cross-sectional online surveys were distributed from December 2023 to January 2024 through a snowball sampling method. Finally, 490 medical students from various local tertiary medical centers were involved. Overall, a majority of the participants showed a positive attitude toward future learning and the usage of AI, manifested as totally willing to acquire relevant knowledge (222/490, 45.3%), totally willing to use AI tools (230/490, 46.9%), and totally desiring that schools would offer AI-related courses (230/490, 46.9%). However, there is still a large proportion (392/490, 80.0%) of participants who expressed concerns regarding ethical issues. The findings also indicated that gender and educational level were significantly correlated with the AI application. Specifically, regression analysis indicated that male participants were more inclined to acquire AI information through social media (odds ratio 0.458, 95% CI 0.33-0.67; P<.001) and that male or graduate-level participants were more likely to use AI for academic writing purposes (odds ratio 0.476, 95% CI 0.38-0.82; P=.001 for male; odds ratio 1.552, 95% CI 1.32-1.77; P=.009 for graduate students, respectively).
Conclusions: Our findings indicate that a general awareness of AI's role in medical education is evident among students. However, subgroup-specific differences must be taken into account, particularly when designing and optimizing educational strategies integrated with AI. This consideration is critical to ensuring that such tools align with the diverse learning needs of distinct student groups.
背景:尽管人工智能(AI)正在迅速融入医学教育,但医学生的见解,特别是在中国背景下,仍然有限。目的:本研究旨在探讨中国医学生对人工智能融入医学教育的认知和态度,以及可能影响其观点的因素。本研究的发现提供了宝贵的见解,以协助医学教育工作者有效地实施这些创新的教育方法。方法:根据目标院校临床医学生总人数估算,计算样本量为379人。采用基于网络的问卷调查方法,调查医学生对人工智能应用的接受程度。该问卷由4个维度的14个问题组成,包括人口特征、对人工智能应用的看法、意愿和担忧。每个维度包含3到4个问题。描述性统计用于将每个变量的频率制成表格。采用卡方检验和多元回归分析测定影响因素。结果:从2023年12月至2024年1月,采用滚雪球抽样法,共发放了566份横断面在线调查。最后,来自各地方三级医疗中心的490名医学生参与了研究。总体而言,大多数参与者对未来学习和使用人工智能持积极态度,表现为非常愿意获取相关知识(222/490,45.3%),非常愿意使用人工智能工具(230/490,46.9%),非常希望学校开设人工智能相关课程(230/490,46.9%)。然而,仍然有很大比例(392/490,80.0%)的参与者表达了对伦理问题的担忧。研究结果还表明,性别和教育水平与人工智能应用显著相关。具体而言,回归分析显示,男性参与者更倾向于通过社交媒体获取人工智能信息(优势比0.458,95% CI 0.33-0.67);结论:我们的研究结果表明,学生普遍意识到人工智能在医学教育中的作用。然而,必须考虑到特定群体的差异,特别是在设计和优化与人工智能相结合的教育策略时。这一考虑对于确保这些工具与不同学生群体的不同学习需求保持一致至关重要。
{"title":"Perceptions and Attitudes of Medical Students Toward the Integration of Large Language Models in Medical Education: Cross-Sectional Survey in China.","authors":"Cheng Zhao, Weiqian Yan, Long Wang, Jing Wu, Herve Pasteur Ndikuriyo, Renhe Yu","doi":"10.2196/66381","DOIUrl":"10.2196/66381","url":null,"abstract":"<p><strong>Background: </strong>Although artificial intelligence (AI) is being rapidly integrated into medical education, insights from medical students, particularly in the Chinese context, remain limited.</p><p><strong>Objective: </strong>This study was designed to explore Chinese medical students' perceptions of and attitudes toward the integration of AI into medical education, as well as the factors that may influence their perspectives. The findings of this research offer valuable insights to assist medical educators in the effective implementation of these innovative educational approaches.</p><p><strong>Methods: </strong>On the basis of the estimated total number of clinical medical students at the target institutions, the sample size was calculated to be 379. A web-based questionnaire survey was designed to investigate the acceptance level of medical students toward the application of AI. The questionnaire consisted of 14 questions across 4 dimensions, which included demographic characteristics, perceptions of AI application, willingness, and concerns. Each dimension contained 3 to 4 questions. Descriptive statistics were used to tabulate the frequency of each variable. Chi-square tests and multiple regression analyses were conducted to measure the influencing factors.</p><p><strong>Results: </strong>A total of 566 cross-sectional online surveys were distributed from December 2023 to January 2024 through a snowball sampling method. Finally, 490 medical students from various local tertiary medical centers were involved. Overall, a majority of the participants showed a positive attitude toward future learning and the usage of AI, manifested as totally willing to acquire relevant knowledge (222/490, 45.3%), totally willing to use AI tools (230/490, 46.9%), and totally desiring that schools would offer AI-related courses (230/490, 46.9%). However, there is still a large proportion (392/490, 80.0%) of participants who expressed concerns regarding ethical issues. The findings also indicated that gender and educational level were significantly correlated with the AI application. Specifically, regression analysis indicated that male participants were more inclined to acquire AI information through social media (odds ratio 0.458, 95% CI 0.33-0.67; P<.001) and that male or graduate-level participants were more likely to use AI for academic writing purposes (odds ratio 0.476, 95% CI 0.38-0.82; P=.001 for male; odds ratio 1.552, 95% CI 1.32-1.77; P=.009 for graduate students, respectively).</p><p><strong>Conclusions: </strong>Our findings indicate that a general awareness of AI's role in medical education is evident among students. However, subgroup-specific differences must be taken into account, particularly when designing and optimizing educational strategies integrated with AI. This consideration is critical to ensuring that such tools align with the diverse learning needs of distinct student groups.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e66381"},"PeriodicalIF":3.2,"publicationDate":"2026-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12978898/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147436295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Andy Cheuk Nam Hwang, Rahul Singh, Elizabeth Ann Barrett, Peng Cao, Varut Vardhanabhuti, Pauline Yeung Ng, Gordon Tin Chun Wong, Michael Tiong Hong Co, Elaine Yuen-Phin Lee
Background: Point-of-care ultrasound training is being increasingly integrated into undergraduate medical education, leading to a substantial demand for trained faculty to provide instruction and feedback.
Objective: This study aimed to develop an adjunct tool, a deep learning-based feedback model, to facilitate student learning.
Methods: Renal ultrasound images (N=2807) were used to train a cascaded deep learning-based feedback model that classified images into three categories: optimal, suboptimal, and incorrect. Suboptimal images were further subcategorized as images with artifact, incorrect gain, and/or incorrect positioning. The model was deployed among year 5 medical students receiving bedside ultrasound training, who were invited to upload renal ultrasound images to an online platform for automated image quality grading and feedback. A mixed methods analysis was used to evaluate students' learning experience. Focus group interviews were organized to qualitatively analyze the successes and challenges of implementation. Quantitative analysis was based on responses to a 5-point Likert scale questionnaire and performance on the objective structured clinical examination (OSCE). Objective structured clinical examination scores were compared with mean OSCE scores from the 2 years preceding implementation of the deep learning-based feedback model.
Results: Focus group interviews identified that the deep learning-based feedback model encouraged self-regulated learning but also recognized that discordant curricular design and hardware limitations impeded its use. The 11-item online questionnaire had a response rate of 42.4% (98/231 students). Among respondents, 32% (31/98) to 48% (47/98) found the model helpful in assisting ultrasound training (Likert score of 4-5 for items 1-3), while 49% (48/98) to 76% (74/98) were satisfied with its usability and their interaction with the model (Likert score of 4-5 for items 4-11). The mean OSCE score was 9.73 (SD 0.76) out of 10, compared with mean scores of 9.35 (SD 1.03; P=.06) and 9.45 (SD 0.97; P=.15) out of 10 in the 2 individual years preceding implementation of the model.
Conclusions: A cascaded deep learning-based feedback model was developed and deployed among year 5 medical students receiving bedside ultrasound training, with positive learner responses and enhanced self-regulated learning. The innovation was associated with increased student engagement and improved ultrasound skill acquisition among novice learners.
{"title":"Development of a Deep Learning-Based Feedback Model to Assist Medical Students Learning Renal Ultrasound Acquisition: Mixed Methods Study.","authors":"Andy Cheuk Nam Hwang, Rahul Singh, Elizabeth Ann Barrett, Peng Cao, Varut Vardhanabhuti, Pauline Yeung Ng, Gordon Tin Chun Wong, Michael Tiong Hong Co, Elaine Yuen-Phin Lee","doi":"10.2196/72110","DOIUrl":"10.2196/72110","url":null,"abstract":"<p><strong>Background: </strong>Point-of-care ultrasound training is being increasingly integrated into undergraduate medical education, leading to a substantial demand for trained faculty to provide instruction and feedback.</p><p><strong>Objective: </strong>This study aimed to develop an adjunct tool, a deep learning-based feedback model, to facilitate student learning.</p><p><strong>Methods: </strong>Renal ultrasound images (N=2807) were used to train a cascaded deep learning-based feedback model that classified images into three categories: optimal, suboptimal, and incorrect. Suboptimal images were further subcategorized as images with artifact, incorrect gain, and/or incorrect positioning. The model was deployed among year 5 medical students receiving bedside ultrasound training, who were invited to upload renal ultrasound images to an online platform for automated image quality grading and feedback. A mixed methods analysis was used to evaluate students' learning experience. Focus group interviews were organized to qualitatively analyze the successes and challenges of implementation. Quantitative analysis was based on responses to a 5-point Likert scale questionnaire and performance on the objective structured clinical examination (OSCE). Objective structured clinical examination scores were compared with mean OSCE scores from the 2 years preceding implementation of the deep learning-based feedback model.</p><p><strong>Results: </strong>Focus group interviews identified that the deep learning-based feedback model encouraged self-regulated learning but also recognized that discordant curricular design and hardware limitations impeded its use. The 11-item online questionnaire had a response rate of 42.4% (98/231 students). Among respondents, 32% (31/98) to 48% (47/98) found the model helpful in assisting ultrasound training (Likert score of 4-5 for items 1-3), while 49% (48/98) to 76% (74/98) were satisfied with its usability and their interaction with the model (Likert score of 4-5 for items 4-11). The mean OSCE score was 9.73 (SD 0.76) out of 10, compared with mean scores of 9.35 (SD 1.03; P=.06) and 9.45 (SD 0.97; P=.15) out of 10 in the 2 individual years preceding implementation of the model.</p><p><strong>Conclusions: </strong>A cascaded deep learning-based feedback model was developed and deployed among year 5 medical students receiving bedside ultrasound training, with positive learner responses and enhanced self-regulated learning. The innovation was associated with increased student engagement and improved ultrasound skill acquisition among novice learners.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e72110"},"PeriodicalIF":3.2,"publicationDate":"2026-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12978925/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147436254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Tianyu Terry Zhang, Rob Saunders, Stephen Pilling, Ciarán O'Driscoll
Background: Cognitive behavioral therapy (CBT) training faces significant challenges, including supervised practice with diverse cases, inconsistent feedback, resource-intensive supervision, and difficulties standardizing competence assessment.
Objective: This study evaluated the acceptability and feasibility of CBT Trainer (TTZ), the first virtual patient platform to provide real-time feedback aligned with established competence frameworks. The mobile app trains psychological practitioners using standardized artificial intelligence patient interactions and the evaluation of therapist responses against competence frameworks to enable structured skill development in a controlled, repeatable environment that complements traditional training methods.
Methods: This mixed methods pilot study used a 2-stage approach. Stage 1 involved usability testing with 4 participants. Stage 2 included 59 participants from psychological practitioner training programs (a Low Intensity CBT Interventions Program and a Doctorate in Clinical Psychology) who engaged with the CBT Trainer voluntarily for over 1 month. Measures of impact included the System Usability Scale (SUS), platform naturalistic engagement, poststudy questionnaire on perceived competency development, comparative evaluation against traditional role-play, and qualitative feedback.
Results: Participants engaged voluntarily with the platform for an average of 95.24 (SD 134.58; median 45.34, IQR 11.57-105.15) minutes of active role-play. Platform usability was rated as excellent (mean SUS 82.20, SD 12.93). Self-reported competence improvement improved most in assessment skills (96.7%) and information gathering (66.7%). When compared to traditional peer role-play exercises, participants rated CBT Trainer moderately favorably (mean 5.90/10, SD 1.94). Qualitative feedback highlighted strengths in competency-aligned feedback, on-demand accessibility, and a psychologically safe practice space.
Conclusions: This pilot study provides evidence that an artificial intelligence-based patient simulation shows promise as a supplementary training tool for psychological therapists who use CBT in their practice, particularly regarding accessibility and immediate feedback. Future research should use randomized controlled designs with objective competence assessments.
{"title":"An AI-Driven Virtual Patient Platform (CBT Trainer) for Training Cognitive Behavioral Therapy Practitioners Against Competencies: Mixed Methods Pilot Study.","authors":"Tianyu Terry Zhang, Rob Saunders, Stephen Pilling, Ciarán O'Driscoll","doi":"10.2196/84091","DOIUrl":"10.2196/84091","url":null,"abstract":"<p><strong>Background: </strong>Cognitive behavioral therapy (CBT) training faces significant challenges, including supervised practice with diverse cases, inconsistent feedback, resource-intensive supervision, and difficulties standardizing competence assessment.</p><p><strong>Objective: </strong>This study evaluated the acceptability and feasibility of CBT Trainer (TTZ), the first virtual patient platform to provide real-time feedback aligned with established competence frameworks. The mobile app trains psychological practitioners using standardized artificial intelligence patient interactions and the evaluation of therapist responses against competence frameworks to enable structured skill development in a controlled, repeatable environment that complements traditional training methods.</p><p><strong>Methods: </strong>This mixed methods pilot study used a 2-stage approach. Stage 1 involved usability testing with 4 participants. Stage 2 included 59 participants from psychological practitioner training programs (a Low Intensity CBT Interventions Program and a Doctorate in Clinical Psychology) who engaged with the CBT Trainer voluntarily for over 1 month. Measures of impact included the System Usability Scale (SUS), platform naturalistic engagement, poststudy questionnaire on perceived competency development, comparative evaluation against traditional role-play, and qualitative feedback.</p><p><strong>Results: </strong>Participants engaged voluntarily with the platform for an average of 95.24 (SD 134.58; median 45.34, IQR 11.57-105.15) minutes of active role-play. Platform usability was rated as excellent (mean SUS 82.20, SD 12.93). Self-reported competence improvement improved most in assessment skills (96.7%) and information gathering (66.7%). When compared to traditional peer role-play exercises, participants rated CBT Trainer moderately favorably (mean 5.90/10, SD 1.94). Qualitative feedback highlighted strengths in competency-aligned feedback, on-demand accessibility, and a psychologically safe practice space.</p><p><strong>Conclusions: </strong>This pilot study provides evidence that an artificial intelligence-based patient simulation shows promise as a supplementary training tool for psychological therapists who use CBT in their practice, particularly regarding accessibility and immediate feedback. Future research should use randomized controlled designs with objective competence assessments.</p>","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e84091"},"PeriodicalIF":3.2,"publicationDate":"2026-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12978919/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147436073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Livia Maria Strasser, Wilma Anschuetz, Fabio Dennstädt, Janna Hastings
<p><strong>Background: </strong>Artificial intelligence continues to transform health care, offering promising applications in clinical practice and medical education. While large language models (LLMs), as a form of generative artificial intelligence, have shown potential to match or surpass medical students in licensing examinations, their performance varies across languages. Recent studies highlight the complex influence and interdependency of factors such as language and model type on LLMs' accuracy; yet, cross-language comparisons remain underexplored.</p><p><strong>Objective: </strong>This study evaluates the performance of LLMs in answering medical multiple-choice questions quantitatively and qualitatively across 3 languages (German, French, and Italian), aiming to uncover model capabilities in a multilingual medical education context.</p><p><strong>Methods: </strong>For this mixed methods study, 114 publicly accessible multiple-choice questions in German, French, and Italian from an online self-assessment tool were analyzed. A quantitative performance analysis of several LLMs developed by OpenAI, Meta AI, Anthropic, and DeepSeek was conducted to evaluate their performance on answering the questions in text-only format. For the comparative analysis, a variation of input question language (German, French, and Italian) and prompt language (English vs language-matched) was used. The 2 best-performing LLMs were then prompted to provide answer explanations for incorrectly answered questions. A subsequent qualitative analysis was conducted on these explanations to identify the reasons leading to the incorrect answers.</p><p><strong>Results: </strong>The performance of LLMs in answering medical multiple-choice questions varied by model and language, showing substantial differences in accuracy (between 64% and 87%). The effect of input question language was significant (P<.01) with models performing best on German questions. Across the analyzed LLMs, prompting in English generally led to better performance in comparison to language-matched prompts, but the top-performing models exceptionally showed comparable results for language-matched prompts. Qualitative analysis revealed that answer explanations of the analyzed models (GPT4o and Claude-Sonnet-3.7) showed different reasoning errors. In several explanations, this occurred despite factual accuracy on the represented topic. Furthermore, this analysis revealed 3 questions to be insufficiently precise.</p><p><strong>Conclusions: </strong>Our results underline the potential of LLMs in answering medical examination questions and highlight the importance of careful consideration of model choice, prompt, and input languages, because of relevant performance variability across these factors. Analysis of answer explanations demonstrates a valuable use case of LLMs for improving examination question quality in medical education, if data security regulations permit their use. Human oversight of language-sen
{"title":"Performance Evaluation of Large Language Models in Multilingual Medical Multiple-Choice Questions: Mixed Methods Study.","authors":"Livia Maria Strasser, Wilma Anschuetz, Fabio Dennstädt, Janna Hastings","doi":"10.2196/81399","DOIUrl":"10.2196/81399","url":null,"abstract":"<p><strong>Background: </strong>Artificial intelligence continues to transform health care, offering promising applications in clinical practice and medical education. While large language models (LLMs), as a form of generative artificial intelligence, have shown potential to match or surpass medical students in licensing examinations, their performance varies across languages. Recent studies highlight the complex influence and interdependency of factors such as language and model type on LLMs' accuracy; yet, cross-language comparisons remain underexplored.</p><p><strong>Objective: </strong>This study evaluates the performance of LLMs in answering medical multiple-choice questions quantitatively and qualitatively across 3 languages (German, French, and Italian), aiming to uncover model capabilities in a multilingual medical education context.</p><p><strong>Methods: </strong>For this mixed methods study, 114 publicly accessible multiple-choice questions in German, French, and Italian from an online self-assessment tool were analyzed. A quantitative performance analysis of several LLMs developed by OpenAI, Meta AI, Anthropic, and DeepSeek was conducted to evaluate their performance on answering the questions in text-only format. For the comparative analysis, a variation of input question language (German, French, and Italian) and prompt language (English vs language-matched) was used. The 2 best-performing LLMs were then prompted to provide answer explanations for incorrectly answered questions. A subsequent qualitative analysis was conducted on these explanations to identify the reasons leading to the incorrect answers.</p><p><strong>Results: </strong>The performance of LLMs in answering medical multiple-choice questions varied by model and language, showing substantial differences in accuracy (between 64% and 87%). The effect of input question language was significant (P<.01) with models performing best on German questions. Across the analyzed LLMs, prompting in English generally led to better performance in comparison to language-matched prompts, but the top-performing models exceptionally showed comparable results for language-matched prompts. Qualitative analysis revealed that answer explanations of the analyzed models (GPT4o and Claude-Sonnet-3.7) showed different reasoning errors. In several explanations, this occurred despite factual accuracy on the represented topic. Furthermore, this analysis revealed 3 questions to be insufficiently precise.</p><p><strong>Conclusions: </strong>Our results underline the potential of LLMs in answering medical examination questions and highlight the importance of careful consideration of model choice, prompt, and input languages, because of relevant performance variability across these factors. Analysis of answer explanations demonstrates a valuable use case of LLMs for improving examination question quality in medical education, if data security regulations permit their use. Human oversight of language-sen","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e81399"},"PeriodicalIF":3.2,"publicationDate":"2026-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12978932/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147436305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
<p><strong>Background: </strong>Interprofessional education (IPE) is a key strategy for enhancing collaboration and patient safety. While evidence for student populations is abundant, studies focusing on licensed physical therapists (PTs), occupational therapists (OTs), and speech-language pathologists (SLPs) remain limited. In contemporary rehabilitation practice, continuous IPE is increasingly important to address professional burnout and the growing complexity of patient needs.</p><p><strong>Objective: </strong>This scoping review aimed to systematically map and synthesize the educational formats, content domains, and reported outcomes of IPE programs specifically targeting licensed PTs, OTs, and SLPs.</p><p><strong>Methods: </strong>Following Joanna Briggs Institute and PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines, we searched the PubMed, Web of Science, Cumulative Index to Nursing and Allied Health Literature, and Educational Resources Information Center databases through December 31, 2025. The eligibility criteria were based on the population, concept, and context framework, including peer-reviewed, English-language studies of licensed PTs, OTs, and SLPs (population) participating in structured IPE interventions (concept) within clinical or community settings (context). Studies focusing solely on students or prelicensure trainees were excluded. Following the screening of 3234 records by independent pairs of reviewers, 9 studies were ultimately selected for inclusion. Methodological quality was appraised using Joanna Briggs Institute critical appraisal checklists and the Mixed Methods Appraisal Tool. Data were synthesized using an evidence gap map to visualize research density across domains relative to established competency frameworks.</p><p><strong>Results: </strong>A total of 9 studies from Australia, the United States, Canada, and the Philippines were included, with sample sizes ranging from 8 to 197. Most used single-group pre-post or mixed methods designs; notably, no randomized controlled trials were identified. Methodological quality was generally high, though limited by the lack of control groups. Systematic mapping identified 7 educational formats, with lectures and discussions being the most dominant across all competency domains. Primary content domains included communication and role clarification. Specific successful interventions included pharmacist-led medication safety workshops and the Kawa model for team building. While participants reported immediate improvements in role understanding and collaborative confidence, simulation-based training showed inconsistent effects on long-term clinical behavior. A substantial evidence gap was identified in experiential learning approaches targeting collaborative leadership.</p><p><strong>Conclusions: </strong>This scoping review adds a novel perspective by focusing exclusively on licensed rehabilitation profes
背景:跨专业教育(IPE)是加强合作和患者安全的关键策略。虽然针对学生群体的证据非常丰富,但针对有执照的物理治疗师(PTs)、职业治疗师(OTs)和语言病理学家(slp)的研究仍然有限。在当代康复实践中,持续的IPE对于解决职业倦怠和患者需求日益复杂的问题越来越重要。目的:本范围综述旨在系统地绘制和综合IPE项目的教育格式、内容领域和报告结果,特别是针对获得许可的PTs、OTs和slp。方法:按照Joanna Briggs研究所和PRISMA-ScR(首选报告项目为系统评价和荟萃分析扩展范围评价)指南,我们检索PubMed, Web of Science,护理和相关健康文献累积索引和教育资源信息中心数据库,截至2025年12月31日。资格标准基于人群、概念和环境框架,包括同行评议的、有执照的PTs、OTs和slp(人群)在临床或社区环境(环境)中参与结构化IPE干预(概念)的英语研究。仅针对学生或执照前受训人员的研究被排除在外。在独立的审稿人对3234条记录进行筛选后,最终选择了9项研究纳入。使用乔安娜布里格斯研究所关键评估清单和混合方法评估工具评估方法学质量。使用证据差距图来合成数据,以可视化相对于已建立的能力框架的跨领域研究密度。结果:共纳入来自澳大利亚、美国、加拿大和菲律宾的9项研究,样本量从8 ~ 197人不等。多采用单组pre-post或混合方法设计;值得注意的是,没有发现随机对照试验。方法质量总体较高,但由于缺乏对照组而受到限制。系统映射确定了7种教育形式,讲座和讨论是所有能力领域中最主要的。主要内容领域包括沟通和角色澄清。具体成功的干预措施包括药剂师领导的用药安全讲习班和Kawa团队建设模式。虽然参与者报告在角色理解和协作信心方面的即时改善,但基于模拟的培训对长期临床行为的影响并不一致。在针对协作领导的体验式学习方法中,发现了实质性的证据差距。结论:本综述通过专门关注有执照的康复专业人员(PTs, ot和slp)增加了一个新的视角,突出了与未获得执照的学生不同的学习需求。它使该领域对潜在的“领导差距”和目前对有经验的临床医生过度依赖教学方法有了更清晰的理解。现实世界的影响表明,卫生保健机构需要向包含客观行为评估的系统的、实践集成的IPE模型过渡。通过纵向项目解决协作领导和团队运作方面的差距,医疗机构可能有助于建立更具弹性的团队文化,最终提高患者安全和康复护理的质量。
{"title":"Educational Formats and Content Domains of Interprofessional Education for Licensed Rehabilitation Professionals: Scoping Review.","authors":"Kohei Ikeda, Takao Kaneko, Someka Hijikuro, Natsuki Inoue, Takuto Nakamura, Taisei Takeda, Junya Uchida, Hirofumi Nagayama","doi":"10.2196/76189","DOIUrl":"10.2196/76189","url":null,"abstract":"<p><strong>Background: </strong>Interprofessional education (IPE) is a key strategy for enhancing collaboration and patient safety. While evidence for student populations is abundant, studies focusing on licensed physical therapists (PTs), occupational therapists (OTs), and speech-language pathologists (SLPs) remain limited. In contemporary rehabilitation practice, continuous IPE is increasingly important to address professional burnout and the growing complexity of patient needs.</p><p><strong>Objective: </strong>This scoping review aimed to systematically map and synthesize the educational formats, content domains, and reported outcomes of IPE programs specifically targeting licensed PTs, OTs, and SLPs.</p><p><strong>Methods: </strong>Following Joanna Briggs Institute and PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines, we searched the PubMed, Web of Science, Cumulative Index to Nursing and Allied Health Literature, and Educational Resources Information Center databases through December 31, 2025. The eligibility criteria were based on the population, concept, and context framework, including peer-reviewed, English-language studies of licensed PTs, OTs, and SLPs (population) participating in structured IPE interventions (concept) within clinical or community settings (context). Studies focusing solely on students or prelicensure trainees were excluded. Following the screening of 3234 records by independent pairs of reviewers, 9 studies were ultimately selected for inclusion. Methodological quality was appraised using Joanna Briggs Institute critical appraisal checklists and the Mixed Methods Appraisal Tool. Data were synthesized using an evidence gap map to visualize research density across domains relative to established competency frameworks.</p><p><strong>Results: </strong>A total of 9 studies from Australia, the United States, Canada, and the Philippines were included, with sample sizes ranging from 8 to 197. Most used single-group pre-post or mixed methods designs; notably, no randomized controlled trials were identified. Methodological quality was generally high, though limited by the lack of control groups. Systematic mapping identified 7 educational formats, with lectures and discussions being the most dominant across all competency domains. Primary content domains included communication and role clarification. Specific successful interventions included pharmacist-led medication safety workshops and the Kawa model for team building. While participants reported immediate improvements in role understanding and collaborative confidence, simulation-based training showed inconsistent effects on long-term clinical behavior. A substantial evidence gap was identified in experiential learning approaches targeting collaborative leadership.</p><p><strong>Conclusions: </strong>This scoping review adds a novel perspective by focusing exclusively on licensed rehabilitation profes","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e76189"},"PeriodicalIF":3.2,"publicationDate":"2026-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12978893/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147436246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Chun Jin Marcus Tan, Wei Wei Dayna Yong, Hui'En Hazel Anne Lin, Jaslyn Oh, How Sheng Rubin Yong, Fang Mei Jayme Khew, Liang Shen, Yujia Gao, Wei Chieh Alfred Kow, Yih Chung Tham, Dianbo Liu, Ching-Yu Cheng, Kee Yuan Ngiam, Yew Sen Yuen, Ray Manotosh, Eng Tat Khoo, Teck Chang Victor Koh, Woon Teck Clement Tan
<p><strong>Background: </strong>Mixed reality has the potential to transform delivery of medical education. With tools such as HoloLens 2, educators can create immersive, interactive simulations that enable students to practice and engage with real-world scenarios in a controlled environment.</p><p><strong>Objective: </strong>We postulated that a hybrid ophthalmology curriculum incorporating EyelearnMR (a simulation application) would be noninferior to traditional teaching. We compared learning outcomes and obtained user feedback.</p><p><strong>Methods: </strong>This was a single-blind, cluster-randomized prospective study. Fourth-year medical students were organized into batches and then assigned to 2 groups: EyelearnMR and control arms. We used a quasi-randomized design with alternation allocation based on clinical grouping. The intervention group had an additional 2 hours of practice with the EyelearnMR devices. During the second week of their posting, a video assessment (5 scenarios with 17 multiple-choice questions) was conducted for both groups-mid-posting for the intervention group and at the end of the posting for the control group. The rationale for assessing the intervention group earlier, in addition to setting a higher bar for EyelearnMR, was to allow for the provision of outcomes showing noninferiority between both groups. In the event of noninferiority, we could demonstrate that EyelearnMR can replace some degree of traditional clinical teaching, even with a shorter total clinical exposure time. Students in the control group were allowed to experience the Eyelearn MR modules for 2 hours at the end of the posting. Both groups were asked to complete the User Experience Questionnaire.</p><p><strong>Results: </strong>This study was funded in February 2023, and recruitment took place from July 2023 to January 2024. A total of 54 students were recruited-24 (44.4%) in the control arm and 30 (55.6%) in the EyelearnMR arm. The EyelearnMR group performed significantly better than the control group (median scores of 16, IQR 15-17, and 15, IQR 14-15, respectively; P=.03; Mann-Whitney U test). A total of 100% (30/30) of the students in the EyelearnMR group scored full marks (3/3) for the technique portion, compared to 70.8% (17/24) of the students in the control group (P=.002). There was no statistically significant difference between the groups for the examination (P=.13) and pathology (P=.33) portions. This was despite the EyelearnMR group having a reduced clinical time of 7 days compared to 10 days in the control group. The User Experience Questionnaire showed positive evaluations for attractiveness (mean 1.413, SD 0.969), efficiency (mean 0.822, SD 1.068), dependability (mean 1.087, SD 0.801), stimulation (mean 1.577, SD 0.845), and novelty (mean 1.606, SD 0.967).</p><p><strong>Conclusions: </strong>EyelearnMR with traditional teaching was noninferior to traditional teaching alone. It provided a comparable experience and supported learning o
{"title":"Application of Mixed Reality for Ophthalmic Clinical Skills and Diagnosis: Prospective Study.","authors":"Chun Jin Marcus Tan, Wei Wei Dayna Yong, Hui'En Hazel Anne Lin, Jaslyn Oh, How Sheng Rubin Yong, Fang Mei Jayme Khew, Liang Shen, Yujia Gao, Wei Chieh Alfred Kow, Yih Chung Tham, Dianbo Liu, Ching-Yu Cheng, Kee Yuan Ngiam, Yew Sen Yuen, Ray Manotosh, Eng Tat Khoo, Teck Chang Victor Koh, Woon Teck Clement Tan","doi":"10.2196/71338","DOIUrl":"10.2196/71338","url":null,"abstract":"<p><strong>Background: </strong>Mixed reality has the potential to transform delivery of medical education. With tools such as HoloLens 2, educators can create immersive, interactive simulations that enable students to practice and engage with real-world scenarios in a controlled environment.</p><p><strong>Objective: </strong>We postulated that a hybrid ophthalmology curriculum incorporating EyelearnMR (a simulation application) would be noninferior to traditional teaching. We compared learning outcomes and obtained user feedback.</p><p><strong>Methods: </strong>This was a single-blind, cluster-randomized prospective study. Fourth-year medical students were organized into batches and then assigned to 2 groups: EyelearnMR and control arms. We used a quasi-randomized design with alternation allocation based on clinical grouping. The intervention group had an additional 2 hours of practice with the EyelearnMR devices. During the second week of their posting, a video assessment (5 scenarios with 17 multiple-choice questions) was conducted for both groups-mid-posting for the intervention group and at the end of the posting for the control group. The rationale for assessing the intervention group earlier, in addition to setting a higher bar for EyelearnMR, was to allow for the provision of outcomes showing noninferiority between both groups. In the event of noninferiority, we could demonstrate that EyelearnMR can replace some degree of traditional clinical teaching, even with a shorter total clinical exposure time. Students in the control group were allowed to experience the Eyelearn MR modules for 2 hours at the end of the posting. Both groups were asked to complete the User Experience Questionnaire.</p><p><strong>Results: </strong>This study was funded in February 2023, and recruitment took place from July 2023 to January 2024. A total of 54 students were recruited-24 (44.4%) in the control arm and 30 (55.6%) in the EyelearnMR arm. The EyelearnMR group performed significantly better than the control group (median scores of 16, IQR 15-17, and 15, IQR 14-15, respectively; P=.03; Mann-Whitney U test). A total of 100% (30/30) of the students in the EyelearnMR group scored full marks (3/3) for the technique portion, compared to 70.8% (17/24) of the students in the control group (P=.002). There was no statistically significant difference between the groups for the examination (P=.13) and pathology (P=.33) portions. This was despite the EyelearnMR group having a reduced clinical time of 7 days compared to 10 days in the control group. The User Experience Questionnaire showed positive evaluations for attractiveness (mean 1.413, SD 0.969), efficiency (mean 0.822, SD 1.068), dependability (mean 1.087, SD 0.801), stimulation (mean 1.577, SD 0.845), and novelty (mean 1.606, SD 0.967).</p><p><strong>Conclusions: </strong>EyelearnMR with traditional teaching was noninferior to traditional teaching alone. It provided a comparable experience and supported learning o","PeriodicalId":36236,"journal":{"name":"JMIR Medical Education","volume":"12 ","pages":"e71338"},"PeriodicalIF":3.2,"publicationDate":"2026-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12978906/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147436083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}