Effectiveness of ChatGPT-4o in developing continuing professional development plans for graduate radiographers: a descriptive study.

IF 9.3 Q1 EDUCATION, SCIENTIFIC DISCIPLINES Journal of Educational Evaluation for Health Professions Pub Date : 2024-01-01 Epub Date: 2024-11-18 DOI:10.3352/jeehp.2024.21.34
Minh Chau, Elio Stefan Arruzza, Kelly Spuur
{"title":"Effectiveness of ChatGPT-4o in developing continuing professional development plans for graduate radiographers: a descriptive study.","authors":"Minh Chau, Elio Stefan Arruzza, Kelly Spuur","doi":"10.3352/jeehp.2024.21.34","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>This study evaluates the use of ChatGPT-4o in creating tailored continuing professional development (CPD) plans for radiography students, addressing the challenge of aligning CPD with Medical Radiation Practice Board of Australia (MRPBA) requirements. We hypothesized that ChatGPT-4o could support students in CPD planning while meeting regulatory standards.</p><p><strong>Methods: </strong>A descriptive, experimental design was used to generate 3 unique CPD plans using ChatGPT-4o, each tailored to hypothetical graduate radiographers in varied clinical settings. Each plan followed MRPBA guidelines, focusing on computed tomography specialization by the second year. Three MRPBA-registered academics assessed the plans using criteria of appropriateness, timeliness, relevance, reflection, and completeness from October 2024 to November 2024. Ratings underwent analysis using the Friedman test and intraclass correlation coefficient (ICC) to measure consistency among evaluators.</p><p><strong>Results: </strong>ChatGPT-4o generated CPD plans generally adhered to regulatory standards across scenarios. The Friedman test indicated no significant differences among raters (P=0.420, 0.761, and 0.807 for each scenario), suggesting consistent scores within scenarios. However, ICC values were low (-0.96, 0.41, and 0.058 for scenarios 1, 2, and 3), revealing variability among raters, particularly in timeliness and completeness criteria, suggesting limitations in the ChatGPT-4o's ability to address individualized and context-specific needs.</p><p><strong>Conclusion: </strong>ChatGPT-4o demonstrates the potential to ease the cognitive demands of CPD planning, offering structured support in CPD development. However, human oversight remains essential to ensure plans are contextually relevant and deeply reflective. Future research should focus on enhancing artificial intelligence's personalization for CPD evaluation, highlighting ChatGPT-4o's potential and limitations as a tool in professional education.</p>","PeriodicalId":46098,"journal":{"name":"Journal of Educational Evaluation for Health Professions","volume":"21 ","pages":"34"},"PeriodicalIF":9.3000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Educational Evaluation for Health Professions","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3352/jeehp.2024.21.34","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/11/18 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"EDUCATION, SCIENTIFIC DISCIPLINES","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose: This study evaluates the use of ChatGPT-4o in creating tailored continuing professional development (CPD) plans for radiography students, addressing the challenge of aligning CPD with Medical Radiation Practice Board of Australia (MRPBA) requirements. We hypothesized that ChatGPT-4o could support students in CPD planning while meeting regulatory standards.

Methods: A descriptive, experimental design was used to generate 3 unique CPD plans using ChatGPT-4o, each tailored to hypothetical graduate radiographers in varied clinical settings. Each plan followed MRPBA guidelines, focusing on computed tomography specialization by the second year. Three MRPBA-registered academics assessed the plans using criteria of appropriateness, timeliness, relevance, reflection, and completeness from October 2024 to November 2024. Ratings underwent analysis using the Friedman test and intraclass correlation coefficient (ICC) to measure consistency among evaluators.

Results: ChatGPT-4o generated CPD plans generally adhered to regulatory standards across scenarios. The Friedman test indicated no significant differences among raters (P=0.420, 0.761, and 0.807 for each scenario), suggesting consistent scores within scenarios. However, ICC values were low (-0.96, 0.41, and 0.058 for scenarios 1, 2, and 3), revealing variability among raters, particularly in timeliness and completeness criteria, suggesting limitations in the ChatGPT-4o's ability to address individualized and context-specific needs.

Conclusion: ChatGPT-4o demonstrates the potential to ease the cognitive demands of CPD planning, offering structured support in CPD development. However, human oversight remains essential to ensure plans are contextually relevant and deeply reflective. Future research should focus on enhancing artificial intelligence's personalization for CPD evaluation, highlighting ChatGPT-4o's potential and limitations as a tool in professional education.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
ChatGPT-4o 在制定放射技师毕业生继续职业发展计划方面的有效性:一项描述性研究。
目的:本研究评估了 ChatGPT-4o 在为放射照相专业学生量身定制持续专业发展(CPD)计划中的应用,以解决持续专业发展与澳大利亚医疗放射实践委员会(MRPBA)要求相一致的难题。我们假设 ChatGPT-4o 可以帮助学生制定 CPD 计划,同时满足监管标准:我们采用描述性实验设计,使用 ChatGPT-4o 生成了 3 个独特的 CPD 计划,每个计划都是为不同临床环境中的假定放射技师毕业生量身定制的。每个计划都遵循 MRPBA 的指导方针,重点是第二年的计算机断层扫描专业化。从 2024 年 10 月到 2024 年 11 月,三位在 MRPBA 注册的学者使用适当性、及时性、相关性、反思性和完整性标准对这些计划进行了评估。评分采用弗里德曼检验和类内相关系数(ICC)进行分析,以衡量评估者之间的一致性:结果:ChatGPT-4o 生成的 CPD 计划在各种情况下总体上符合监管标准。弗里德曼检验结果表明,评价者之间没有明显差异(每个方案的 P=0.420、0.761 和 0.807),这表明方案内的评分是一致的。然而,ICC 值较低(情景 1、2 和 3 的 ICC 值分别为-0.96、0.41 和 0.058),显示出评分者之间的差异,尤其是在及时性和完整性标准方面,这表明 ChatGPT-4o 在满足个性化和特定环境需求方面存在局限性:结论:ChatGPT-4o 展示了缓解持续专业发展规划认知需求的潜力,为持续专业发展提供了结构化支持。不过,要确保计划与实际情况相关并具有深刻的反思性,人工监督仍然必不可少。未来的研究应侧重于加强人工智能在持续专业发展评估方面的个性化,突出 ChatGPT-4o 作为专业教育工具的潜力和局限性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
9.60
自引率
9.10%
发文量
32
审稿时长
5 weeks
期刊介绍: Journal of Educational Evaluation for Health Professions aims to provide readers the state-of-the art practical information on the educational evaluation for health professions so that to increase the quality of undergraduate, graduate, and continuing education. It is specialized in educational evaluation including adoption of measurement theory to medical health education, promotion of high stakes examination such as national licensing examinations, improvement of nationwide or international programs of education, computer-based testing, computerized adaptive testing, and medical health regulatory bodies. Its field comprises a variety of professions that address public medical health as following but not limited to: Care workers Dental hygienists Dental technicians Dentists Dietitians Emergency medical technicians Health educators Medical record technicians Medical technologists Midwives Nurses Nursing aides Occupational therapists Opticians Oriental medical doctors Oriental medicine dispensers Oriental pharmacists Pharmacists Physical therapists Physicians Prosthetists and Orthotists Radiological technologists Rehabilitation counselor Sanitary technicians Speech-language therapists.
期刊最新文献
Importance, performance frequency, and predicted future importance of dietitians’ jobs by practicing dietitians in Korea: a survey study Presidential address 2024: the expansion of computer-based testing to numerous health professions licensing examinations in Korea, preparation of computer-based practical tests, and adoption of the medical metaverse. Development and validity evidence for the resident-led large group teaching assessment instrument in the United States: a methodological study. ChatGPT (GPT-4) passed the Japanese National License Examination for Pharmacists in 2022, answering all items including those with diagrams: a descriptive study. Discovering social learning ecosystems during clinical clerkship from United States medical students’ feedback encounters: a content analysis.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1