首页 > 最新文献

Journal of Educational Evaluation for Health Professions最新文献

英文 中文
Training satisfaction and future employment consideration among physician and nursing trainees at rural Veterans Affairs facilities in the United States during COVID-19: a time-series before and after study 在 COVID-19 期间,美国农村退伍军人事务机构的医生和护理受训人员对培训的满意度和未来就业考虑:一项前后时间序列研究。
IF 9.3 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-09-24 DOI: 10.3352/jeehp.2024.21.25
Heather Northcraft, Tiffany Radcliff, Anne Reid Griffin, Jia Bai, Aram Dobalian

Purpose: The coronavirus disease 2019 (COVID-19) pandemic limited healthcare professional education and training opportunities in rural communities. Because the US Department of Veterans Affairs (VA) has robust programs to train clinicians in the United States, this study examined VA trainee perspectives regarding pandemic-related training in rural and urban areas and interest in future employment with the VA.

Methods: Survey responses were collected nationally from VA physicians and nursing trainees before and after COVID-19 (2018 to 2021). Logistic regression models were used to test the association between pandemic timing (pre-pandemic or pandemic), trainee program (physician or nurse), and the interaction of trainee pandemic timing and program on VA trainee satisfaction and trainee likelihood to consider future VA employment in rural and urban areas.

Results: While physician trainees at urban facilities reported decreases in overall training satisfaction and corresponding decreases in the likelihood of considering future VA employment from pre-pandemic to pandemic, rural physician trainees showed no changes in either outcome. In contrast, while nursing trainees at both urban and rural sites had decreases in training satisfaction associated with the pandemic, there was no corresponding effect on the likelihood of future employment by nurses at either urban or rural VA sites.

Conclusion: The study’s findings suggest differences in the training experiences of physicians and nurses at rural sites, as well as between physician trainees at urban and rural sites. Understanding these nuances can inform the development of targeted approaches to address the ongoing provider shortages that rural communities in the United States are facing.

目的:COVID-19 大流行限制了农村社区的医疗保健专业教育和培训机会。由于美国退伍军人事务部(VA)拥有在美国培训临床医生的强大项目,本研究考察了退伍军人事务部受训人员对城乡地区大流行相关培训的看法以及未来在退伍军人事务部就业的兴趣:在 COVID-19 前后(2018 年至 2021 年),在全国范围内收集了退伍军人事务部医生和护理受训人员的调查回复。使用逻辑回归模型检验了大流行时间(大流行前或大流行)、受训者项目(医生或护士)以及受训者大流行时间和项目的交互作用对退伍军人事务部受训者满意度和受训者考虑未来在农村和城市地区退伍军人事务部就业的可能性之间的关联:结果:从大流行前到大流行期间,城市医疗机构的医生受训人员对培训的总体满意度有所下降,考虑将来在退伍军人事务部就业的可能性也相应下降,而农村的医生受训人员在这两项结果上都没有变化。与此相反,虽然城市和农村医疗机构的护士学员的培训满意度都因大流行而下降,但对城市或农村退伍军人医疗机构护士未来就业的可能性却没有相应的影响:研究结果表明,农村医疗点的医生和护士以及城市和农村医疗点的受训医生在培训经历上存在差异。了解这些细微差别可以为制定有针对性的方法提供信息,以解决美国农村社区持续面临的医疗服务提供者短缺问题。
{"title":"Training satisfaction and future employment consideration among physician and nursing trainees at rural Veterans Affairs facilities in the United States during COVID-19: a time-series before and after study","authors":"Heather Northcraft, Tiffany Radcliff, Anne Reid Griffin, Jia Bai, Aram Dobalian","doi":"10.3352/jeehp.2024.21.25","DOIUrl":"10.3352/jeehp.2024.21.25","url":null,"abstract":"<p><strong>Purpose: </strong>The coronavirus disease 2019 (COVID-19) pandemic limited healthcare professional education and training opportunities in rural communities. Because the US Department of Veterans Affairs (VA) has robust programs to train clinicians in the United States, this study examined VA trainee perspectives regarding pandemic-related training in rural and urban areas and interest in future employment with the VA.</p><p><strong>Methods: </strong>Survey responses were collected nationally from VA physicians and nursing trainees before and after COVID-19 (2018 to 2021). Logistic regression models were used to test the association between pandemic timing (pre-pandemic or pandemic), trainee program (physician or nurse), and the interaction of trainee pandemic timing and program on VA trainee satisfaction and trainee likelihood to consider future VA employment in rural and urban areas.</p><p><strong>Results: </strong>While physician trainees at urban facilities reported decreases in overall training satisfaction and corresponding decreases in the likelihood of considering future VA employment from pre-pandemic to pandemic, rural physician trainees showed no changes in either outcome. In contrast, while nursing trainees at both urban and rural sites had decreases in training satisfaction associated with the pandemic, there was no corresponding effect on the likelihood of future employment by nurses at either urban or rural VA sites.</p><p><strong>Conclusion: </strong>The study’s findings suggest differences in the training experiences of physicians and nurses at rural sites, as well as between physician trainees at urban and rural sites. Understanding these nuances can inform the development of targeted approaches to address the ongoing provider shortages that rural communities in the United States are facing.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"25"},"PeriodicalIF":9.3,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11528153/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142356105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A new performance evaluation indicator for the LEE Jong-wook Fellowship Program of Korea Foundation for International Healthcare to better assess its long-term educational impacts: a Delphi study. 韩国国际医疗基金会李钟郁奖学金项目的新绩效评估指标,以更好地评估其长期教育影响:德尔菲研究。
IF 9.3 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-10-02 DOI: 10.3352/jeehp.2024.21.27
Minkyung Oh, Bo Young Yoon

Purpose: The Dr. LEE Jong-wook Fellowship Program, established by the Korea Foundation for International Healthcare (KOFIH), aims to strengthen healthcare capacity in partner countries. The aim of the study was to develop new performance evaluation indicators for the program to better assess long-term educational impact across various courses and professional roles.

Methods: A 3-stage process was employed. First, a literature review of established evaluation models (Kirkpatrick’s 4 levels, context/input/process/product evaluation model, Organization for Economic Cooperation and Development Assistance Committee criteria) was conducted to devise evaluation criteria. Second, these criteria were validated via a 2-round Delphi survey with 18 experts in training projects from May 2021 to June 2021. Third, the relative importance of the evaluation criteria was determined using the analytic hierarchy process (AHP), calculating weights and ensuring consistency through the consistency index and consistency ratio (CR), with CR values below 0.1 indicating acceptable consistency.

Results: The literature review led to a combined evaluation model, resulting in 4 evaluation areas, 20 items, and 92 indicators. The Delphi surveys confirmed the validity of these indicators, with content validity ratio values exceeding 0.444. The AHP analysis assigned weights to each indicator, and CR values below 0.1 indicated consistency. The final set of evaluation indicators was confirmed through a workshop with KOFIH and adopted as the new evaluation tool.

Conclusion: The developed evaluation framework provides a comprehensive tool for assessing the long-term outcomes of the Dr. LEE Jong-wook Fellowship Program. It enhances evaluation capabilities and supports improvements in the training program’s effectiveness and international healthcare collaboration.

目的:由韩国国际医疗保健基金会(KOFIH)设立的李钟郁博士奖学金项目旨在加强伙伴国的医疗保健能力。本研究旨在为该计划制定新的绩效评估指标,以更好地评估不同课程和专业角色的长期教育影响:方法:采用了三阶段流程。首先,对已有的评估模式(Kirkpatrick 的 4 个等级、CIPP 模式、经合组织 DAC 标准)进行文献综述,以制定评估标准。其次,在 2021 年 5 月至 2021 年 6 月期间,对 18 名培训项目专家进行了两轮德尔菲调查,对这些标准进行了验证。第三,使用层次分析法(AHP)确定评价标准的相对重要性,计算权重,并通过一致性指数(CI)和一致性比率(CR)确保一致性,CR 值低于 0.1 表示一致性可接受:文献综述得出了一个综合评价模型,包括 4 个评价领域、20 个项目和 92 个指标。德尔菲调查证实了这些指标的有效性,其内容效度比值超过了 0.444。AHP 分析为每个指标分配了权重,CR 值低于 0.1 表示一致性。通过与 KFIH 举办研讨会,最终确认了这套评估指标,并将其作为新的评估工具:制定的评估框架为评估李钟郁博士奖学金项目的长期成果提供了一个全面的工具。它增强了评估能力,有助于提高培训计划的有效性和国际医疗合作。
{"title":"A new performance evaluation indicator for the LEE Jong-wook Fellowship Program of Korea Foundation for International Healthcare to better assess its long-term educational impacts: a Delphi study.","authors":"Minkyung Oh, Bo Young Yoon","doi":"10.3352/jeehp.2024.21.27","DOIUrl":"10.3352/jeehp.2024.21.27","url":null,"abstract":"<p><strong>Purpose: </strong>The Dr. LEE Jong-wook Fellowship Program, established by the Korea Foundation for International Healthcare (KOFIH), aims to strengthen healthcare capacity in partner countries. The aim of the study was to develop new performance evaluation indicators for the program to better assess long-term educational impact across various courses and professional roles.</p><p><strong>Methods: </strong>A 3-stage process was employed. First, a literature review of established evaluation models (Kirkpatrick’s 4 levels, context/input/process/product evaluation model, Organization for Economic Cooperation and Development Assistance Committee criteria) was conducted to devise evaluation criteria. Second, these criteria were validated via a 2-round Delphi survey with 18 experts in training projects from May 2021 to June 2021. Third, the relative importance of the evaluation criteria was determined using the analytic hierarchy process (AHP), calculating weights and ensuring consistency through the consistency index and consistency ratio (CR), with CR values below 0.1 indicating acceptable consistency.</p><p><strong>Results: </strong>The literature review led to a combined evaluation model, resulting in 4 evaluation areas, 20 items, and 92 indicators. The Delphi surveys confirmed the validity of these indicators, with content validity ratio values exceeding 0.444. The AHP analysis assigned weights to each indicator, and CR values below 0.1 indicated consistency. The final set of evaluation indicators was confirmed through a workshop with KOFIH and adopted as the new evaluation tool.</p><p><strong>Conclusion: </strong>The developed evaluation framework provides a comprehensive tool for assessing the long-term outcomes of the Dr. LEE Jong-wook Fellowship Program. It enhances evaluation capabilities and supports improvements in the training program’s effectiveness and international healthcare collaboration.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"27"},"PeriodicalIF":9.3,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11535579/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142366885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Challenges and potential improvements in the Accreditation Standards of the Korean Institute of Medical Education and Evaluation 2019 (ASK2019) derived through meta-evaluation: a cross-sectional study 通过荟萃评估得出的《2019 年韩国医学教育与评估研究院认证标准》(ASK2019)的挑战和潜在改进:一项横断面研究。
IF 4.4 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-04-02 DOI: 10.3352/jeehp.2024.21.8
Yoonjung Lee, Min-jung Lee, Junmoo Ahn, Chungwon Ha, Ye Ji Kang, Cheol Woong Jung, Dong-Mi Yoo, Jihye Yu, Seung-Hee Lee

Purpose: This study aimed to identify challenges and potential improvements in Korea’s medical education accreditation process according to the Accreditation Standards of the Korean Institute of Medical Education and Evaluation 2019 (ASK2019). Meta-evaluation was conducted to survey the experiences and perceptions of stakeholders, including self-assessment committee members, site visit committee members, administrative staff, and medical school professors.

Methods: A cross-sectional study was conducted using surveys sent to 40 medical schools. The 332 participants included self-assessment committee members, site visit team members, administrative staff, and medical school professors. The t-test, one-way analysis of variance and the chi-square test were used to analyze and compare opinions on medical education accreditation between the categories of participants.

Results: Site visit committee members placed greater importance on the necessity of accreditation than faculty members. A shared positive view on accreditation’s role in improving educational quality was seen among self-evaluation committee members and professors. Administrative staff highly regarded the Korean Institute of Medical Education and Evaluation’s reliability and objectivity, unlike the self-evaluation committee members. Site visit committee members positively perceived the clarity of accreditation standards, differing from self-assessment committee members. Administrative staff were most optimistic about implementing standards. However, the accreditation process encountered challenges, especially in duplicating content and preparing self-evaluation reports. Finally, perceptions regarding the accuracy of final site visit reports varied significantly between the self-evaluation committee members and the site visit committee members.

Conclusion: This study revealed diverse views on medical education accreditation, highlighting the need for improved communication, expectation alignment, and stakeholder collaboration to refine the accreditation process and quality.

目的:本研究旨在根据《2019 年韩国医学教育与评估研究院评审标准》(ASK2019),确定韩国医学教育评审过程中的挑战和潜在改进措施。通过元评价,调查了包括自评委员会成员、现场考察委员会成员、行政人员和医学教授在内的利益相关者的经验和看法:向 40 所医学院校发送了调查问卷,开展了一项横断面研究。332名参与者包括自评委员会成员、现场考察小组成员、行政人员和医学院教授。采用 t 检验、单因素方差分析和卡方检验来分析和比较各类参与者对医学教育评审的意见:结果:现场考察委员会成员比教职员工更重视评审的必要性。自我评估委员会成员和教授都对评审在提高教育质量方面的作用持积极态度。与自我评估委员会成员不同,行政人员高度评价韩国医学教育与评价院的可靠性和客观性。实地考察评估人员对评审标准的清晰度持肯定态度,这一点与自我评估委员会成员不同。行政人员对标准的实施最为乐观。然而,评审过程遇到了挑战,特别是在重复内容和编写自我评估报告方面。最后,自我评估委员会成员和现场考察委员会成员对最终现场考察报告准确性的看法存在很大差异:本研究揭示了对医学教育评审的不同看法,强调了加强沟通、调整期望值和利益相关者合作以完善评审过程和提高评审质量的必要性。
{"title":"Challenges and potential improvements in the Accreditation Standards of the Korean Institute of Medical Education and Evaluation 2019 (ASK2019) derived through meta-evaluation: a cross-sectional study","authors":"Yoonjung Lee, Min-jung Lee, Junmoo Ahn, Chungwon Ha, Ye Ji Kang, Cheol Woong Jung, Dong-Mi Yoo, Jihye Yu, Seung-Hee Lee","doi":"10.3352/jeehp.2024.21.8","DOIUrl":"10.3352/jeehp.2024.21.8","url":null,"abstract":"<p><strong>Purpose: </strong>This study aimed to identify challenges and potential improvements in Korea’s medical education accreditation process according to the Accreditation Standards of the Korean Institute of Medical Education and Evaluation 2019 (ASK2019). Meta-evaluation was conducted to survey the experiences and perceptions of stakeholders, including self-assessment committee members, site visit committee members, administrative staff, and medical school professors.</p><p><strong>Methods: </strong>A cross-sectional study was conducted using surveys sent to 40 medical schools. The 332 participants included self-assessment committee members, site visit team members, administrative staff, and medical school professors. The t-test, one-way analysis of variance and the chi-square test were used to analyze and compare opinions on medical education accreditation between the categories of participants.</p><p><strong>Results: </strong>Site visit committee members placed greater importance on the necessity of accreditation than faculty members. A shared positive view on accreditation’s role in improving educational quality was seen among self-evaluation committee members and professors. Administrative staff highly regarded the Korean Institute of Medical Education and Evaluation’s reliability and objectivity, unlike the self-evaluation committee members. Site visit committee members positively perceived the clarity of accreditation standards, differing from self-assessment committee members. Administrative staff were most optimistic about implementing standards. However, the accreditation process encountered challenges, especially in duplicating content and preparing self-evaluation reports. Finally, perceptions regarding the accuracy of final site visit reports varied significantly between the self-evaluation committee members and the site visit committee members.</p><p><strong>Conclusion: </strong>This study revealed diverse views on medical education accreditation, highlighting the need for improved communication, expectation alignment, and stakeholder collaboration to refine the accreditation process and quality.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"8"},"PeriodicalIF":4.4,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11108703/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140337062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Performance of GPT-3.5 and GPT-4 on standardized urology knowledge assessment items in the United States: a descriptive study. 美国 GPT-3.5 和 GPT-4 在标准化泌尿科知识评估项目上的表现:一项描述性研究。
IF 9.3 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-07-08 DOI: 10.3352/jeehp.2024.21.17
Max Samuel Yudovich, Elizaveta Makarova, Christian Michael Hague, Jay Dilip Raman

Purpose: This study aimed to evaluate the performance of Chat Generative Pre-Trained Transformer (ChatGPT) with respect to standardized urology multiple-choice items in the United States.

Methods: In total, 700 multiple-choice urology board exam-style items were submitted to GPT-3.5 and GPT-4, and responses were recorded. Items were categorized based on topic and question complexity (recall, interpretation, and problem-solving). The accuracy of GPT-3.5 and GPT-4 was compared across item types in February 2024.

Results: GPT-4 answered 44.4% of items correctly compared to 30.9% for GPT-3.5 (P>0.0001). GPT-4 (vs. GPT-3.5) had higher accuracy with urologic oncology (43.8% vs. 33.9%, P=0.03), sexual medicine (44.3% vs. 27.8%, P=0.046), and pediatric urology (47.1% vs. 27.1%, P=0.012) items. Endourology (38.0% vs. 25.7%, P=0.15), reconstruction and trauma (29.0% vs. 21.0%, P=0.41), and neurourology (49.0% vs. 33.3%, P=0.11) items did not show significant differences in performance across versions. GPT-4 also outperformed GPT-3.5 with respect to recall (45.9% vs. 27.4%, P<0.00001), interpretation (45.6% vs. 31.5%, P=0.0005), and problem-solving (41.8% vs. 34.5%, P=0.56) type items. This difference was not significant for the higher-complexity items.

Conclusion: s: ChatGPT performs relatively poorly on standardized multiple-choice urology board exam-style items, with GPT-4 outperforming GPT-3.5. The accuracy was below the proposed minimum passing standards for the American Board of Urology's Continuing Urologic Certification knowledge reinforcement activity (60%). As artificial intelligence progresses in complexity, ChatGPT may become more capable and accurate with respect to board examination items. For now, its responses should be scrutinized.

目的:本研究旨在评估 Chat Generative Pre-Trained Transformer(ChatGPT)在美国标准化泌尿科选择题方面的性能:共向 GPT-3.5 和 GPT-4 提交了 700 个泌尿外科委员会考试类型的多项选择题,并记录了答案。根据题目和问题复杂程度(回忆、解释和解决问题)对项目进行分类。2024 年 2 月,比较了 GPT-3.5 和 GPT-4 在不同项目类型中的准确性:结果:GPT-4 回答正确率为 44.4%,而 GPT-3.5 为 30.9%(P>0.0001)。GPT-4(vs.GPT-3.5)在泌尿肿瘤学(43.8% vs. 33.9%,P=0.03)、性医学(44.3% vs. 27.8%,P=0.046)和小儿泌尿学(47.1% vs. 27.1%,P=0.012)项目上的准确率更高。内泌尿学(38.0% vs. 25.7%,P=0.15)、重建与创伤(29.0% vs. 21.0%,P=0.41)和神经泌尿学(49.0% vs. 33.3%,P=0.11)项目在不同版本中的表现没有显著差异。在回忆率方面,GPT-4 也优于 GPT-3.5(45.9% 对 27.4%,P=0.41):ChatGPT 在标准化的泌尿外科医师资格考试多选题上的表现相对较差,GPT-4 的表现优于 GPT-3.5。准确率低于美国泌尿外科委员会泌尿外科继续认证知识强化活动的最低合格标准(60%)。随着人工智能复杂性的提高,ChatGPT 在委员会考试项目方面的能力和准确性可能会越来越高。就目前而言,应该对它的回答进行仔细检查。
{"title":"Performance of GPT-3.5 and GPT-4 on standardized urology knowledge assessment items in the United States: a descriptive study.","authors":"Max Samuel Yudovich, Elizaveta Makarova, Christian Michael Hague, Jay Dilip Raman","doi":"10.3352/jeehp.2024.21.17","DOIUrl":"https://doi.org/10.3352/jeehp.2024.21.17","url":null,"abstract":"<p><strong>Purpose: </strong>This study aimed to evaluate the performance of Chat Generative Pre-Trained Transformer (ChatGPT) with respect to standardized urology multiple-choice items in the United States.</p><p><strong>Methods: </strong>In total, 700 multiple-choice urology board exam-style items were submitted to GPT-3.5 and GPT-4, and responses were recorded. Items were categorized based on topic and question complexity (recall, interpretation, and problem-solving). The accuracy of GPT-3.5 and GPT-4 was compared across item types in February 2024.</p><p><strong>Results: </strong>GPT-4 answered 44.4% of items correctly compared to 30.9% for GPT-3.5 (P>0.0001). GPT-4 (vs. GPT-3.5) had higher accuracy with urologic oncology (43.8% vs. 33.9%, P=0.03), sexual medicine (44.3% vs. 27.8%, P=0.046), and pediatric urology (47.1% vs. 27.1%, P=0.012) items. Endourology (38.0% vs. 25.7%, P=0.15), reconstruction and trauma (29.0% vs. 21.0%, P=0.41), and neurourology (49.0% vs. 33.3%, P=0.11) items did not show significant differences in performance across versions. GPT-4 also outperformed GPT-3.5 with respect to recall (45.9% vs. 27.4%, P<0.00001), interpretation (45.6% vs. 31.5%, P=0.0005), and problem-solving (41.8% vs. 34.5%, P=0.56) type items. This difference was not significant for the higher-complexity items.</p><p><strong>Conclusion: </strong>s: ChatGPT performs relatively poorly on standardized multiple-choice urology board exam-style items, with GPT-4 outperforming GPT-3.5. The accuracy was below the proposed minimum passing standards for the American Board of Urology's Continuing Urologic Certification knowledge reinforcement activity (60%). As artificial intelligence progresses in complexity, ChatGPT may become more capable and accurate with respect to board examination items. For now, its responses should be scrutinized.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"17"},"PeriodicalIF":9.3,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141560038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The effect of simulation-based training on problem-solving skills, critical thinking skills, and self-efficacy among nursing students in Vietnam: a before-and-after study. 模拟训练对越南护理专业学生解决问题能力、批判性思维能力和自我效能感的影响:一项前后对比研究。
IF 9.3 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-09-23 DOI: 10.3352/jeehp.2024.21.24
Tran Thi Hoang Oanh, Luu Thi Thuy, Ngo Thi Thu Huyen

Purpose: This study investigated the effect of simulation-based training on nursing students’ problem-solving skills, critical thinking skills, and self-efficacy.

Methods: A single-group pretest and posttest study was conducted among 173 second-year nursing students at a public university in Vietnam from May 2021 to July 2022. Each student participated in the adult nursing preclinical practice course, which utilized a moderate-fidelity simulation teaching approach. Instruments including the Personal Problem-Solving Inventory Scale, Critical Thinking Skills Questionnaire, and General Self-Efficacy Questionnaire were employed to measure participants’ problem-solving skills, critical thinking skills, and self-efficacy. Data were analyzed using descriptive statistics and the paired-sample t-test with the significance level set at P<0.05.

Results: The mean score of the Personal Problem-Solving Inventory posttest (127.24±12.11) was lower than the pretest score (131.42±16.95), suggesting an improvement in the problem-solving skills of the participants (t172 =2.55, P=0.011). There was no statistically significant difference in critical thinking skills between the pretest and posttest (P=0.854). Self-efficacy among nursing students showed a substantial increase from the pretest (27.91±5.26) to the posttest (28.71±3.81), with t172 =-2.26 and P=0.025.

Conclusion: The results suggest that simulation-based training can improve problem-solving skills and increase self-efficacy among nursing students. Therefore, the integration of simulation-based training in nursing education is recommended.

目的:本研究探讨了模拟训练对护理专业学生解决问题的能力、批判性思维能力和自我效能感的影响:方法:2021 年 5 月至 2022 年 7 月,在越南一所公立大学的 173 名护理专业二年级学生中开展了一项单组前测后测研究。每位学生都参加了成人护理临床前实践课程,该课程采用了中度保真模拟教学法。研究采用了个人问题解决量表、批判性思维能力问卷和一般自我效能感问卷等工具来测量参与者的问题解决能力、批判性思维能力和自我效能感。数据分析采用描述性统计和配对样本 t 检验,显著性水平为 PResults:个人问题解决量表》后测平均分(127.24±12.11)低于前测平均分(131.42±16.95),表明参与者的问题解决能力有所提高(t172=2.55,P=0.011)。批判性思维能力在前测和后测之间没有统计学差异(P=0.854)。护生的自我效能感从前测(27.91±5.26)到后测(28.71±3.81)有大幅提高,t172=-2.26,P=0.025:结果表明,模拟训练可以提高护生解决问题的能力,增强自我效能感。因此,建议在护理教学中融入模拟训练。
{"title":"The effect of simulation-based training on problem-solving skills, critical thinking skills, and self-efficacy among nursing students in Vietnam: a before-and-after study.","authors":"Tran Thi Hoang Oanh, Luu Thi Thuy, Ngo Thi Thu Huyen","doi":"10.3352/jeehp.2024.21.24","DOIUrl":"10.3352/jeehp.2024.21.24","url":null,"abstract":"<p><strong>Purpose: </strong>This study investigated the effect of simulation-based training on nursing students’ problem-solving skills, critical thinking skills, and self-efficacy.</p><p><strong>Methods: </strong>A single-group pretest and posttest study was conducted among 173 second-year nursing students at a public university in Vietnam from May 2021 to July 2022. Each student participated in the adult nursing preclinical practice course, which utilized a moderate-fidelity simulation teaching approach. Instruments including the Personal Problem-Solving Inventory Scale, Critical Thinking Skills Questionnaire, and General Self-Efficacy Questionnaire were employed to measure participants’ problem-solving skills, critical thinking skills, and self-efficacy. Data were analyzed using descriptive statistics and the paired-sample t-test with the significance level set at P<0.05.</p><p><strong>Results: </strong>The mean score of the Personal Problem-Solving Inventory posttest (127.24±12.11) was lower than the pretest score (131.42±16.95), suggesting an improvement in the problem-solving skills of the participants (t172 =2.55, P=0.011). There was no statistically significant difference in critical thinking skills between the pretest and posttest (P=0.854). Self-efficacy among nursing students showed a substantial increase from the pretest (27.91±5.26) to the posttest (28.71±3.81), with t172 =-2.26 and P=0.025.</p><p><strong>Conclusion: </strong>The results suggest that simulation-based training can improve problem-solving skills and increase self-efficacy among nursing students. Therefore, the integration of simulation-based training in nursing education is recommended.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"24"},"PeriodicalIF":9.3,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11480641/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142298256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Opportunities, challenges, and future directions of large language models, including ChatGPT in medical education: a systematic scoping review 包括 ChatGPT 在内的大型语言模型在医学教育中的机遇、挑战和未来发展方向:系统性范围审查。
IF 4.4 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-03-15 DOI: 10.3352/jeehp.2024.21.6
Xiaojun Xu, Yixiao Chen, Jing Miao

Background: ChatGPT is a large language model (LLM) based on artificial intelligence (AI) capable of responding in multiple languages and generating nuanced and highly complex responses. While ChatGPT holds promising applications in medical education, its limitations and potential risks cannot be ignored.

Methods: A scoping review was conducted for English articles discussing ChatGPT in the context of medical education published after 2022. A literature search was performed using PubMed/MEDLINE, Embase, and Web of Science databases, and information was extracted from the relevant studies that were ultimately included.

Results: ChatGPT exhibits various potential applications in medical education, such as providing personalized learning plans and materials, creating clinical practice simulation scenarios, and assisting in writing articles. However, challenges associated with academic integrity, data accuracy, and potential harm to learning were also highlighted in the literature. The paper emphasizes certain recommendations for using ChatGPT, including the establishment of guidelines. Based on the review, 3 key research areas were proposed: cultivating the ability of medical students to use ChatGPT correctly, integrating ChatGPT into teaching activities and processes, and proposing standards for the use of AI by medical students.

Conclusion: ChatGPT has the potential to transform medical education, but careful consideration is required for its full integration. To harness the full potential of ChatGPT in medical education, attention should not only be given to the capabilities of AI but also to its impact on students and teachers.

背景介绍ChatGPT 是一种基于人工智能 (AI) 的大型语言模型 (LLM),能够以多种语言进行应答,并生成细致入微、高度复杂的应答。虽然 ChatGPT 在医学教育中的应用前景广阔,但其局限性和潜在风险也不容忽视:方法:我们对 2022 年后发表的讨论医学教育背景下 ChatGPT 的英文文章进行了一次范围审查。我们使用 PubMed/MEDLINE、Embase 和 Web of Science 数据库进行了文献检索,并从最终纳入的相关研究中提取了信息:ChatGPT在医学教育中具有多种潜在应用,如提供个性化学习计划和材料、创建临床实践模拟场景以及协助撰写文章。然而,文献中也强调了与学术诚信、数据准确性和对学习的潜在危害相关的挑战。本文强调了使用 ChatGPT 的一些建议,包括制定指导方针。在综述的基础上,提出了 3 个重点研究领域:培养医学生正确使用 ChatGPT 的能力、将 ChatGPT 融入教学活动和过程、提出医学生使用人工智能的标准:结论:ChatGPT 具有改变医学教育的潜力,但在全面整合时需要慎重考虑。要在医学教育中充分发挥 ChatGPT 的潜力,不仅要关注人工智能的能力,还要关注其对学生和教师的影响。
{"title":"Opportunities, challenges, and future directions of large language models, including ChatGPT in medical education: a systematic scoping review","authors":"Xiaojun Xu, Yixiao Chen, Jing Miao","doi":"10.3352/jeehp.2024.21.6","DOIUrl":"10.3352/jeehp.2024.21.6","url":null,"abstract":"<p><strong>Background: </strong>ChatGPT is a large language model (LLM) based on artificial intelligence (AI) capable of responding in multiple languages and generating nuanced and highly complex responses. While ChatGPT holds promising applications in medical education, its limitations and potential risks cannot be ignored.</p><p><strong>Methods: </strong>A scoping review was conducted for English articles discussing ChatGPT in the context of medical education published after 2022. A literature search was performed using PubMed/MEDLINE, Embase, and Web of Science databases, and information was extracted from the relevant studies that were ultimately included.</p><p><strong>Results: </strong>ChatGPT exhibits various potential applications in medical education, such as providing personalized learning plans and materials, creating clinical practice simulation scenarios, and assisting in writing articles. However, challenges associated with academic integrity, data accuracy, and potential harm to learning were also highlighted in the literature. The paper emphasizes certain recommendations for using ChatGPT, including the establishment of guidelines. Based on the review, 3 key research areas were proposed: cultivating the ability of medical students to use ChatGPT correctly, integrating ChatGPT into teaching activities and processes, and proposing standards for the use of AI by medical students.</p><p><strong>Conclusion: </strong>ChatGPT has the potential to transform medical education, but careful consideration is required for its full integration. To harness the full potential of ChatGPT in medical education, attention should not only be given to the capabilities of AI but also to its impact on students and teachers.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"6"},"PeriodicalIF":4.4,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11035906/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140132845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Discovering social learning ecosystems during clinical clerkship from United States medical students’ feedback encounters: a content analysis. 从美国医科学生的反馈遭遇中发现临床实习期间的社会学习生态系统:内容分析。
IF 4.4 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-02-28 DOI: 10.3352/jeehp.2024.21.5
Anna Therese Cianciolo, Heeyoung Han, Lydia Anne Howes, Debra Lee Klamen, Sophia Matos

Purpose: We examined United States medical students’ self-reported feedback encounters during clerkship training to better understand in situ feedback practices. Specifically, we asked: Who do students receive feedback from, about what, when, where, and how do they use it? We explored whether curricular expectations for preceptors’ written commentary aligned with feedback as it occurs naturalistically in the workplace.

Methods: This study occurred from July 2021 to February 2022 at Southern Illinois University School of Medicine. We used qualitative survey-based experience sampling to gather students’ accounts of their feedback encounters in 8 core specialties. We analyzed the who, what, when, where, and why of 267 feedback encounters reported by 11 clerkship students over 30 weeks. Code frequencies were mapped qualitatively to explore patterns in feedback encounters.

Results: Clerkship feedback occurs in patterns apparently related to the nature of clinical work in each specialty. These patterns may be attributable to each specialty’s “social learning ecosystem”—the distinctive learning environment shaped by the social and material aspects of a given specialty’s work, which determine who preceptors are, what students do with preceptors, and what skills or attributes matter enough to preceptors to comment on.

Conclusion: Comprehensive, standardized expectations for written feedback across specialties conflict with the reality of workplace-based learning. Preceptors may be better able—and more motivated—to document student performance that occurs as a natural part of everyday work. Nurturing social learning ecosystems could facilitate workplace-based learning such that, across specialties, students acquire a comprehensive clinical skillset appropriate for graduation.

目的:我们研究了美国医科学生在实习培训期间自我报告的反馈情况,以更好地了解现场反馈实践。具体来说,我们询问学生从谁那里获得反馈,反馈的内容、时间、地点以及他们如何使用反馈?我们探讨了课程对实习指导教师书面评语的期望是否与工作场所自然发生的反馈一致:本研究于 2021 年 7 月至 2022 年 2 月在南伊利诺伊大学医学院进行。我们使用基于定性调查的经验抽样,收集了学生在 8 个核心专业中遇到的反馈情况。我们分析了 11 名实习学生在 30 周内报告的 267 次反馈中的 "谁"、"什么"、"何时"、"何地 "和 "为什么"。我们对代码频率进行了定性映射,以探索反馈遭遇的模式:结果:实习反馈的模式显然与各专业临床工作的性质有关。这些模式可能归因于每个专科的社会学习生态系统--由特定专科工作的社会和物质方面所形成的独特的学习环境,这些方面决定了谁是实习医生、学生与实习医生一起做了什么,以及哪些技能或特质对实习医生来说足够重要,以至于需要进行评论:结论:各专业对书面反馈的全面、标准化要求与基于工作场所的学习现实相冲突。作为日常工作的自然组成部分,实习医生可能更有能力、也更有动力记录学生的表现。培养社会学习生态系统可以促进以工作场所为基础的学习,从而使不同专业的学生获得适合毕业的全面临床技能。
{"title":"Discovering social learning ecosystems during clinical clerkship from United States medical students’ feedback encounters: a content analysis.","authors":"Anna Therese Cianciolo, Heeyoung Han, Lydia Anne Howes, Debra Lee Klamen, Sophia Matos","doi":"10.3352/jeehp.2024.21.5","DOIUrl":"10.3352/jeehp.2024.21.5","url":null,"abstract":"<p><strong>Purpose: </strong>We examined United States medical students’ self-reported feedback encounters during clerkship training to better understand in situ feedback practices. Specifically, we asked: Who do students receive feedback from, about what, when, where, and how do they use it? We explored whether curricular expectations for preceptors’ written commentary aligned with feedback as it occurs naturalistically in the workplace.</p><p><strong>Methods: </strong>This study occurred from July 2021 to February 2022 at Southern Illinois University School of Medicine. We used qualitative survey-based experience sampling to gather students’ accounts of their feedback encounters in 8 core specialties. We analyzed the who, what, when, where, and why of 267 feedback encounters reported by 11 clerkship students over 30 weeks. Code frequencies were mapped qualitatively to explore patterns in feedback encounters.</p><p><strong>Results: </strong>Clerkship feedback occurs in patterns apparently related to the nature of clinical work in each specialty. These patterns may be attributable to each specialty’s “social learning ecosystem”—the distinctive learning environment shaped by the social and material aspects of a given specialty’s work, which determine who preceptors are, what students do with preceptors, and what skills or attributes matter enough to preceptors to comment on.</p><p><strong>Conclusion: </strong>Comprehensive, standardized expectations for written feedback across specialties conflict with the reality of workplace-based learning. Preceptors may be better able—and more motivated—to document student performance that occurs as a natural part of everyday work. Nurturing social learning ecosystems could facilitate workplace-based learning such that, across specialties, students acquire a comprehensive clinical skillset appropriate for graduation.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"5"},"PeriodicalIF":4.4,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10948917/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139984162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparison of virtual and in-person simulations for sepsis and trauma resuscitation training in Singapore: a randomized controlled trial. 新加坡脓毒症和创伤复苏培训中虚拟和现场模拟的比较:随机对照试验。
IF 9.3 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-11-18 DOI: 10.3352/jeehp.2024.21.33
Matthew Jian Wen Low, Gene Wai Han Chan, Zisheng Li, Yiwen Koh, Chi Loong Jen, Zi Yao Lee, Lenard Tai Win Cheng

Purpose: This study aimed to compare cognitive, non-cognitive, and overall learning outcomes for sepsis and trauma resuscitation skills in novices with virtual patient simulation (VPS) versus in-person simulation (IPS).

Methods: A randomized controlled trial was conducted on junior doctors in emergency departments from January to December 2022, comparing 70 minutes of VPS (n=19) versus IPS (n=21) in sepsis and trauma resuscitation. Using the nominal group technique, we created skills assessment checklists and determined Bloom's taxonomy domains for each checklist item. Two blinded raters observed participants leading 1 sepsis and 1 trauma resuscitation simulation. Satisfaction was measured using the Student Satisfaction with Learning Scale (SSLS). The SSLS and checklist scores were analyzed using the 2-tailed t-test.

Results: For sepsis, there was no significant difference between VPS and IPS in overall scores (2.0; 95% confidence interval [CI], -1.4 to 5.4; Cohen's d=0.38), as well as in items that were cognitive (1.1; 95% CI, -1.5 to 3.7) and not only cognitive (0.9; 95% CI, -0.4 to 2.2). Likewise, for trauma, there was no significant difference in overall scores (-0.9; 95% CI, -4.1 to 2.3; Cohen's d=0.19), as well as in items that were cognitive (-0.3; 95% CI, -2.8 to 2.1) and not only cognitive (-0.6; 95% CI, -2.4 to 1.3). The median SSLS scores were lower with VPS than with IPS (-3.0; 95% CI, -1.0 to -5.0).

Conclusion: For novices, there were no major differences in overall and non-cognitive learning outcomes for sepsis and trauma resuscitation between VPS and IPS. Learners were more satisfied with IPS than with VPS (clinicaltrials.gov identifier: NCT05201950).

目的:本研究旨在比较虚拟患者模拟(VPS)与现场模拟(IPS)对新手败血症和创伤复苏技能的认知、非认知和总体学习效果:2022 年 1 月至 12 月期间,我们对急诊科的初级医生进行了一项随机对照试验,比较了 70 分钟 VPS(19 人)与 IPS(21 人)在败血症和创伤复苏方面的效果。我们使用名义小组技术创建了技能评估核对表,并确定了每个核对表项目的布卢姆分类学领域。两名双盲评分员观察了学员在败血症和创伤复苏模拟教学中的表现。满意度采用学生学习满意度量表(SSLS)进行测量。采用双尾 t 检验对 SSLS 和核对表得分进行分析:对于败血症,VPS 和 IPS 在总分(2.0;95% 置信区间[CI],-1.4 至 5.4;Cohen's d=0.38)以及认知项目(1.1;95% 置信区间,-1.5 至 3.7)和非认知项目(0.9;95% 置信区间,-0.4 至 2.2)方面均无显著差异。同样,在创伤方面,总分(-0.9;95% CI,-4.1 至 2.3;Cohen's d=0.19)以及认知项目(-0.3;95% CI,-2.8 至 2.1)和非认知项目(-0.6;95% CI,-2.4 至 1.3)均无显著差异。VPS 的 SSLS 中位数得分低于 IPS(-3.0;95% CI,-1.0 至-5.0):对于新手而言,VPS 和 IPS 在败血症和创伤复苏的总体和非认知学习成果方面没有重大差异。学员对 IPS 的满意度高于 VPS(clinicaltrials.gov identifier: NCT05201950)。
{"title":"Comparison of virtual and in-person simulations for sepsis and trauma resuscitation training in Singapore: a randomized controlled trial.","authors":"Matthew Jian Wen Low, Gene Wai Han Chan, Zisheng Li, Yiwen Koh, Chi Loong Jen, Zi Yao Lee, Lenard Tai Win Cheng","doi":"10.3352/jeehp.2024.21.33","DOIUrl":"10.3352/jeehp.2024.21.33","url":null,"abstract":"<p><strong>Purpose: </strong>This study aimed to compare cognitive, non-cognitive, and overall learning outcomes for sepsis and trauma resuscitation skills in novices with virtual patient simulation (VPS) versus in-person simulation (IPS).</p><p><strong>Methods: </strong>A randomized controlled trial was conducted on junior doctors in emergency departments from January to December 2022, comparing 70 minutes of VPS (n=19) versus IPS (n=21) in sepsis and trauma resuscitation. Using the nominal group technique, we created skills assessment checklists and determined Bloom's taxonomy domains for each checklist item. Two blinded raters observed participants leading 1 sepsis and 1 trauma resuscitation simulation. Satisfaction was measured using the Student Satisfaction with Learning Scale (SSLS). The SSLS and checklist scores were analyzed using the 2-tailed t-test.</p><p><strong>Results: </strong>For sepsis, there was no significant difference between VPS and IPS in overall scores (2.0; 95% confidence interval [CI], -1.4 to 5.4; Cohen's d=0.38), as well as in items that were cognitive (1.1; 95% CI, -1.5 to 3.7) and not only cognitive (0.9; 95% CI, -0.4 to 2.2). Likewise, for trauma, there was no significant difference in overall scores (-0.9; 95% CI, -4.1 to 2.3; Cohen's d=0.19), as well as in items that were cognitive (-0.3; 95% CI, -2.8 to 2.1) and not only cognitive (-0.6; 95% CI, -2.4 to 1.3). The median SSLS scores were lower with VPS than with IPS (-3.0; 95% CI, -1.0 to -5.0).</p><p><strong>Conclusion: </strong>For novices, there were no major differences in overall and non-cognitive learning outcomes for sepsis and trauma resuscitation between VPS and IPS. Learners were more satisfied with IPS than with VPS (clinicaltrials.gov identifier: NCT05201950).</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"33"},"PeriodicalIF":9.3,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142648693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Increased accessibility of computer-based testing for residency application to a hospital in Brazil with item characteristics comparable to paper-based testing: a psychometric study. 提高巴西一家医院住院医师申请计算机测试的可及性,其项目特征与纸质测试相当:心理测量学研究。
IF 9.3 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-11-11 DOI: 10.3352/jeehp.2024.21.32
Marcos Carvalho Borges, Luciane Loures Santos, Paulo Henrique Manso, Elaine Christine Dantas Moisés, Pedro Soler Coltro, Priscilla Costa Fonseca, Paulo Roberto Alves Gentil, Rodrigo de Carvalho Santana, Lucas Faria Rodrigues, Benedito Carlos Maciel, Hilton Marcos Alves Ricz

Purpose: With the COVID-19 pandemic, online high-stakes exams have become a viable alternative. This study evaluated the feasibility of computer-based testing (CBT) for medical residency applications in Brazil and its impacts on item quality and applicants' access compared to paper-based testing.

Methods: In 2020, an online CBT was conducted in a Ribeirao Preto Clinical Hospital in Brazil. In total, 120 multiple-choice question items were constructed. Two years later, the exam was performed as paper-based testing. Item construction processes were similar for both exams. Difficulty and discrimination indexes, point-biserial coefficient, difficulty, discrimination, guessing parameters, and Cronbach's alpha coefficient were measured based on the item response and classical test theories. Internet stability for applicants was monitored.

Results: In 2020, 4,846 individuals (57.1% female, mean age of 26.64 ± 3.37 years) applied to the residency program, versus 2,196 individuals (55.2% female, mean age of 26.47 ± 3.20 years) in 2022. For CBT, there was an increase of 2,650 (120.7%) applicants, albeit with significant differences in demographic characteristics. There was a significant increase in applicants from more distant and lower-income Brazilian regions, such as the North (5.6% vs. 2.7%) and Northeast (16.9% vs. 9.0%). No significant differences were found in difficulty and discrimination indexes, point-biserial coefficients, and Cronbach's alpha coefficients between the 2 exams.

Conclusion: Online CBT with multiple-choice questions was a viable format for a residency application exam, improving accessibility without compromising exam integrity and quality.

目的:随着 COVID-19 的流行,在线高风险考试已成为一种可行的选择。本研究评估了在巴西申请住院医师资格时使用计算机辅助考试(CBT)的可行性,以及与纸质考试相比,计算机辅助考试对项目质量和申请人获得考试机会的影响:2020 年,巴西里贝拉普雷图临床医院开展了在线 CBT。共设计了 120 道选择题。两年后,该考试以纸质测试的形式进行。两次考试的题目构建过程相似。根据项目反应理论和经典测验理论,测量了难度和区分度指数、点-比列系数、难度、区分度、猜测参数和克朗巴赫α系数。对申请人的互联网稳定性进行了监测:2020 年有 4846 人(57.1% 为女性,平均年龄为 26.64 ± 3.37 岁)申请住院医师培训项目,而 2022 年为 2196 人(55.2% 为女性,平均年龄为 26.47 ± 3.20 岁)。CBT 的申请人数增加了 2,650 人(120.7%),但在人口统计学特征方面存在显著差异。来自巴西较远和较低收入地区的申请人明显增加,如北部(5.6% 对 2.7%)和东北部(16.9% 对 9.0%)。两种考试的难度和区分度指数、点-比列系数和克朗巴赫α系数均无明显差异:带有多项选择题的在线 CBT 是一种可行的住院医师申请考试形式,既提高了考试的可及性,又不影响考试的完整性和质量。
{"title":"Increased accessibility of computer-based testing for residency application to a hospital in Brazil with item characteristics comparable to paper-based testing: a psychometric study.","authors":"Marcos Carvalho Borges, Luciane Loures Santos, Paulo Henrique Manso, Elaine Christine Dantas Moisés, Pedro Soler Coltro, Priscilla Costa Fonseca, Paulo Roberto Alves Gentil, Rodrigo de Carvalho Santana, Lucas Faria Rodrigues, Benedito Carlos Maciel, Hilton Marcos Alves Ricz","doi":"10.3352/jeehp.2024.21.32","DOIUrl":"https://doi.org/10.3352/jeehp.2024.21.32","url":null,"abstract":"<p><strong>Purpose: </strong>With the COVID-19 pandemic, online high-stakes exams have become a viable alternative. This study evaluated the feasibility of computer-based testing (CBT) for medical residency applications in Brazil and its impacts on item quality and applicants' access compared to paper-based testing.</p><p><strong>Methods: </strong>In 2020, an online CBT was conducted in a Ribeirao Preto Clinical Hospital in Brazil. In total, 120 multiple-choice question items were constructed. Two years later, the exam was performed as paper-based testing. Item construction processes were similar for both exams. Difficulty and discrimination indexes, point-biserial coefficient, difficulty, discrimination, guessing parameters, and Cronbach's alpha coefficient were measured based on the item response and classical test theories. Internet stability for applicants was monitored.</p><p><strong>Results: </strong>In 2020, 4,846 individuals (57.1% female, mean age of 26.64 ± 3.37 years) applied to the residency program, versus 2,196 individuals (55.2% female, mean age of 26.47 ± 3.20 years) in 2022. For CBT, there was an increase of 2,650 (120.7%) applicants, albeit with significant differences in demographic characteristics. There was a significant increase in applicants from more distant and lower-income Brazilian regions, such as the North (5.6% vs. 2.7%) and Northeast (16.9% vs. 9.0%). No significant differences were found in difficulty and discrimination indexes, point-biserial coefficients, and Cronbach's alpha coefficients between the 2 exams.</p><p><strong>Conclusion: </strong>Online CBT with multiple-choice questions was a viable format for a residency application exam, improving accessibility without compromising exam integrity and quality.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"32"},"PeriodicalIF":9.3,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142630096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ChatGPT (GPT-4) passed the Japanese National License Examination for Pharmacists in 2022, answering all items including those with diagrams: a descriptive study. ChatGPT (GPT-4) 在 2022 年通过了日本全国药剂师执照考试,回答了包括图表在内的所有题目:一项描述性研究。
IF 4.4 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Pub Date : 2024-01-01 Epub Date: 2024-02-28 DOI: 10.3352/jeehp.2024.21.4
Hiroyasu Sato, Katsuhiko Ogasawara

Purpose: The objective of this study was to assess the performance of ChatGPT (GPT-4) on all items, including those with diagrams, in the Japanese National License Examination for Pharmacists (JNLEP) and compare it with the previous GPT-3.5 model’s performance.

Methods: The 107th JNLEP, conducted in 2022, with 344 items input into the GPT-4 model, was targeted for this study. Separately, 284 items, excluding those with diagrams, were entered into the GPT-3.5 model. The answers were categorized and analyzed to determine accuracy rates based on categories, subjects, and presence or absence of diagrams. The accuracy rates were compared to the main passing criteria (overall accuracy rate ≥62.9%).

Results: The overall accuracy rate for all items in the 107th JNLEP in GPT-4 was 72.5%, successfully meeting all the passing criteria. For the set of items without diagrams, the accuracy rate was 80.0%, which was significantly higher than that of the GPT-3.5 model (43.5%). The GPT-4 model demonstrated an accuracy rate of 36.1% for items that included diagrams.

Conclusion: Advancements that allow GPT-4 to process images have made it possible for LLMs to answer all items in medical-related license examinations. This study’s findings confirm that ChatGPT (GPT-4) possesses sufficient knowledge to meet the passing criteria.

目的:本研究旨在评估 ChatGPT(GPT-4)在日本国家执业药师资格考试(JNLEP)中所有项目(包括带图表的项目)上的表现,并将其与之前的 GPT-3.5 模型的表现进行比较:本研究以 2022 年举行的第 107 届日本国家执业药师资格考试(JNLEP)为对象,将 344 个项目输入 GPT-4 模型。另外,在 GPT-3.5 模型中输入了 284 个项目(不包括有图表的项目)。对答案进行了分类和分析,以确定基于类别、主题和有无图表的准确率。准确率与主要及格标准(总准确率≥62.9%)进行了比较:第107届日本语能力考试GPT-4中所有题目的总正确率为72.5%,成功地达到了所有及格标准。对于无图表的项目集,正确率为 80.0%,明显高于 GPT-3.5 模型(43.5%)。对于包含图表的项目,GPT-4 模型的准确率为 36.1%:结论:GPT-4 在处理图像方面的进步使法律硕士有可能回答医学相关执照考试中的所有题目。本研究的结果证实,ChatGPT(GPT-4)拥有足够的知识来满足合格标准。
{"title":"ChatGPT (GPT-4) passed the Japanese National License Examination for Pharmacists in 2022, answering all items including those with diagrams: a descriptive study.","authors":"Hiroyasu Sato, Katsuhiko Ogasawara","doi":"10.3352/jeehp.2024.21.4","DOIUrl":"10.3352/jeehp.2024.21.4","url":null,"abstract":"<p><strong>Purpose: </strong>The objective of this study was to assess the performance of ChatGPT (GPT-4) on all items, including those with diagrams, in the Japanese National License Examination for Pharmacists (JNLEP) and compare it with the previous GPT-3.5 model’s performance.</p><p><strong>Methods: </strong>The 107th JNLEP, conducted in 2022, with 344 items input into the GPT-4 model, was targeted for this study. Separately, 284 items, excluding those with diagrams, were entered into the GPT-3.5 model. The answers were categorized and analyzed to determine accuracy rates based on categories, subjects, and presence or absence of diagrams. The accuracy rates were compared to the main passing criteria (overall accuracy rate ≥62.9%).</p><p><strong>Results: </strong>The overall accuracy rate for all items in the 107th JNLEP in GPT-4 was 72.5%, successfully meeting all the passing criteria. For the set of items without diagrams, the accuracy rate was 80.0%, which was significantly higher than that of the GPT-3.5 model (43.5%). The GPT-4 model demonstrated an accuracy rate of 36.1% for items that included diagrams.</p><p><strong>Conclusion: </strong>Advancements that allow GPT-4 to process images have made it possible for LLMs to answer all items in medical-related license examinations. This study’s findings confirm that ChatGPT (GPT-4) possesses sufficient knowledge to meet the passing criteria.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"4"},"PeriodicalIF":4.4,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10948916/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139984149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Educational Evaluation for Health Professions
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1