比较人工智能学习模型在医学生解答组织学和胚胎学选择题中的表现

IF 2 3区 医学 Q2 ANATOMY & MORPHOLOGY Annals of Anatomy-Anatomischer Anzeiger Pub Date : 2024-03-21 DOI:10.1016/j.aanat.2024.152261
Miloš Bajčetić , Aleksandar Mirčić , Jelena Rakočević, Danilo Đoković, Katarina Milutinović, Ivan Zaletel
{"title":"比较人工智能学习模型在医学生解答组织学和胚胎学选择题中的表现","authors":"Miloš Bajčetić ,&nbsp;Aleksandar Mirčić ,&nbsp;Jelena Rakočević,&nbsp;Danilo Đoković,&nbsp;Katarina Milutinović,&nbsp;Ivan Zaletel","doi":"10.1016/j.aanat.2024.152261","DOIUrl":null,"url":null,"abstract":"<div><h3>Introduction</h3><p>The appearance of artificial intelligence language models (AI LMs) in the form of chatbots has gained a lot of popularity worldwide, potentially interfering with different aspects of education, including medical education as well. The present study aims to assess the accuracy and consistency of different AI LMs regarding the histology and embryology knowledge obtained during the 1st year of medical studies.</p></div><div><h3>Methods</h3><p>Five different chatbots (ChatGPT, Bing AI, Bard AI, Perplexity AI, and ChatSonic) were given two sets of multiple-choice questions (MCQs). AI LMs test results were compared to the same test results obtained from 1st year medical students. Chatbots were instructed to use revised Bloom’s taxonomy when classifying questions depending on hierarchical cognitive domains. Simultaneously, two histology teachers independently rated the questions applying the same criteria, followed by the comparison between chatbots’ and teachers’ question classification. The consistency of chatbots’ answers was explored by giving the chatbots the same tests two months apart.</p></div><div><h3>Results</h3><p>AI LMs successfully and correctly solved MCQs regarding histology and embryology material. All five chatbots showed better results than the 1st year medical students on both histology and embryology tests. Chatbots showed poor results when asked to classify the questions according to revised Bloom’s cognitive taxonomy compared to teachers. There was an inverse correlation between the difficulty of questions and their correct classification by the chatbots. Retesting the chatbots after two months showed a lack of consistency concerning both MCQs answers and question classification according to revised Bloom’s taxonomy learning stage.</p></div><div><h3>Conclusion</h3><p>Despite the ability of certain chatbots to provide correct answers to the majority of diverse and heterogeneous questions, a lack of consistency in answers over time warrants their careful use as a medical education tool.</p></div>","PeriodicalId":50974,"journal":{"name":"Annals of Anatomy-Anatomischer Anzeiger","volume":"254 ","pages":"Article 152261"},"PeriodicalIF":2.0000,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Comparing the performance of artificial intelligence learning models to medical students in solving histology and embryology multiple choice questions\",\"authors\":\"Miloš Bajčetić ,&nbsp;Aleksandar Mirčić ,&nbsp;Jelena Rakočević,&nbsp;Danilo Đoković,&nbsp;Katarina Milutinović,&nbsp;Ivan Zaletel\",\"doi\":\"10.1016/j.aanat.2024.152261\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><h3>Introduction</h3><p>The appearance of artificial intelligence language models (AI LMs) in the form of chatbots has gained a lot of popularity worldwide, potentially interfering with different aspects of education, including medical education as well. The present study aims to assess the accuracy and consistency of different AI LMs regarding the histology and embryology knowledge obtained during the 1st year of medical studies.</p></div><div><h3>Methods</h3><p>Five different chatbots (ChatGPT, Bing AI, Bard AI, Perplexity AI, and ChatSonic) were given two sets of multiple-choice questions (MCQs). AI LMs test results were compared to the same test results obtained from 1st year medical students. Chatbots were instructed to use revised Bloom’s taxonomy when classifying questions depending on hierarchical cognitive domains. Simultaneously, two histology teachers independently rated the questions applying the same criteria, followed by the comparison between chatbots’ and teachers’ question classification. The consistency of chatbots’ answers was explored by giving the chatbots the same tests two months apart.</p></div><div><h3>Results</h3><p>AI LMs successfully and correctly solved MCQs regarding histology and embryology material. All five chatbots showed better results than the 1st year medical students on both histology and embryology tests. Chatbots showed poor results when asked to classify the questions according to revised Bloom’s cognitive taxonomy compared to teachers. There was an inverse correlation between the difficulty of questions and their correct classification by the chatbots. Retesting the chatbots after two months showed a lack of consistency concerning both MCQs answers and question classification according to revised Bloom’s taxonomy learning stage.</p></div><div><h3>Conclusion</h3><p>Despite the ability of certain chatbots to provide correct answers to the majority of diverse and heterogeneous questions, a lack of consistency in answers over time warrants their careful use as a medical education tool.</p></div>\",\"PeriodicalId\":50974,\"journal\":{\"name\":\"Annals of Anatomy-Anatomischer Anzeiger\",\"volume\":\"254 \",\"pages\":\"Article 152261\"},\"PeriodicalIF\":2.0000,\"publicationDate\":\"2024-03-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Annals of Anatomy-Anatomischer Anzeiger\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0940960224000530\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ANATOMY & MORPHOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Annals of Anatomy-Anatomischer Anzeiger","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0940960224000530","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ANATOMY & MORPHOLOGY","Score":null,"Total":0}
引用次数: 0

摘要

引言 以聊天机器人形式出现的人工智能语言模型(AI LMs)在全球范围内大受欢迎,有可能干扰教育的各个方面,包括医学教育。本研究旨在评估不同的人工智能语言模型对医学专业一年级学生所学的组织学和胚胎学知识的准确性和一致性。人工智能 LM 的测试结果与一年级医学生的相同测试结果进行了比较。在根据认知领域的层次对问题进行分类时,指导聊天机器人使用修订后的布鲁姆分类法。同时,两名组织学教师采用相同的标准对问题进行独立评分,然后比较聊天机器人和教师的问题分类。结果AI LMs 成功并正确地解决了有关组织学和胚胎学材料的 MCQ。所有五个聊天机器人在组织学和胚胎学测试中的成绩都优于一年级医学生。与教师相比,当要求聊天机器人根据修订后的布鲁姆认知分类法对问题进行分类时,聊天机器人的成绩较差。问题的难度与聊天机器人的正确分类之间存在反相关关系。两个月后对聊天机器人进行的重新测试表明,MCQ 答案和根据修订版布鲁姆分类法学习阶段进行的问题分类都缺乏一致性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Comparing the performance of artificial intelligence learning models to medical students in solving histology and embryology multiple choice questions

Introduction

The appearance of artificial intelligence language models (AI LMs) in the form of chatbots has gained a lot of popularity worldwide, potentially interfering with different aspects of education, including medical education as well. The present study aims to assess the accuracy and consistency of different AI LMs regarding the histology and embryology knowledge obtained during the 1st year of medical studies.

Methods

Five different chatbots (ChatGPT, Bing AI, Bard AI, Perplexity AI, and ChatSonic) were given two sets of multiple-choice questions (MCQs). AI LMs test results were compared to the same test results obtained from 1st year medical students. Chatbots were instructed to use revised Bloom’s taxonomy when classifying questions depending on hierarchical cognitive domains. Simultaneously, two histology teachers independently rated the questions applying the same criteria, followed by the comparison between chatbots’ and teachers’ question classification. The consistency of chatbots’ answers was explored by giving the chatbots the same tests two months apart.

Results

AI LMs successfully and correctly solved MCQs regarding histology and embryology material. All five chatbots showed better results than the 1st year medical students on both histology and embryology tests. Chatbots showed poor results when asked to classify the questions according to revised Bloom’s cognitive taxonomy compared to teachers. There was an inverse correlation between the difficulty of questions and their correct classification by the chatbots. Retesting the chatbots after two months showed a lack of consistency concerning both MCQs answers and question classification according to revised Bloom’s taxonomy learning stage.

Conclusion

Despite the ability of certain chatbots to provide correct answers to the majority of diverse and heterogeneous questions, a lack of consistency in answers over time warrants their careful use as a medical education tool.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Annals of Anatomy-Anatomischer Anzeiger
Annals of Anatomy-Anatomischer Anzeiger 医学-解剖学与形态学
CiteScore
4.40
自引率
22.70%
发文量
137
审稿时长
33 days
期刊介绍: Annals of Anatomy publish peer reviewed original articles as well as brief review articles. The journal is open to original papers covering a link between anatomy and areas such as •molecular biology, •cell biology •reproductive biology •immunobiology •developmental biology, neurobiology •embryology as well as •neuroanatomy •neuroimmunology •clinical anatomy •comparative anatomy •modern imaging techniques •evolution, and especially also •aging
期刊最新文献
Implementing 3D printing and extended reality in anatomy education: Organization, evolution, and assessment of an innovative teaching program in an undergraduate medical school in the Netherlands Visible Human Project based applications can prompt integrating cross-sectional anatomy into the medical school curriculum when combined with radiological modalities: A three-year cross-sectional observational study Geometric morphometric analysis of plastinated brain sections using computer-based methods: Evaluating shrinkage and shape changes Advances in upper urinary tract anatomy through imaging techniques How to deal with the variability of peripheral nerve lesion patterns after inguinal herniotomy? A descriptive approach for a new terminology in clinical practice
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1