Investigating the comparative superiority of artificial intelligence programs in assessing knowledge levels regarding ocular inflammation, uvea diseases, and treatment modalities.

IF 1 Q4 OPHTHALMOLOGY Taiwan Journal of Ophthalmology Pub Date : 2024-09-13 eCollection Date: 2024-07-01 DOI:10.4103/tjo.TJO-D-23-00166
Eyupcan Sensoy, Mehmet Citirik
{"title":"Investigating the comparative superiority of artificial intelligence programs in assessing knowledge levels regarding ocular inflammation, uvea diseases, and treatment modalities.","authors":"Eyupcan Sensoy, Mehmet Citirik","doi":"10.4103/tjo.TJO-D-23-00166","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>The purpose of the study was to evaluate the knowledge level of the Chat Generative Pretrained Transformer (ChatGPT), Bard, and Bing artificial intelligence (AI) chatbots regarding ocular inflammation, uveal diseases, and treatment modalities, and to investigate their relative performance compared to one another.</p><p><strong>Materials and methods: </strong>Thirty-six questions related to ocular inflammation, uveal diseases, and treatment modalities were posed to the ChatGPT, Bard, and Bing AI chatbots, and both correct and incorrect responses were recorded. The accuracy rates were compared using the Chi-squared test.</p><p><strong>Results: </strong>The ChatGPT provided correct answers to 52.8% of the questions, while Bard answered 38.9% correctly, and Bing answered 44.4% correctly. All three AI programs provided identical responses to 20 (55.6%) of the questions, with 45% of these responses being correct and 55% incorrect. No significant difference was observed between the correct and incorrect responses from the three AI chatbots (<i>P</i> = 0.654).</p><p><strong>Conclusion: </strong>AI chatbots should be developed to provide widespread access to accurate information about ocular inflammation, uveal diseases, and treatment modalities. Future research could explore ways to enhance the performance of these chatbots.</p>","PeriodicalId":44978,"journal":{"name":"Taiwan Journal of Ophthalmology","volume":"14 3","pages":"409-413"},"PeriodicalIF":1.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11488809/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Taiwan Journal of Ophthalmology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.4103/tjo.TJO-D-23-00166","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/7/1 0:00:00","PubModel":"eCollection","JCR":"Q4","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose: The purpose of the study was to evaluate the knowledge level of the Chat Generative Pretrained Transformer (ChatGPT), Bard, and Bing artificial intelligence (AI) chatbots regarding ocular inflammation, uveal diseases, and treatment modalities, and to investigate their relative performance compared to one another.

Materials and methods: Thirty-six questions related to ocular inflammation, uveal diseases, and treatment modalities were posed to the ChatGPT, Bard, and Bing AI chatbots, and both correct and incorrect responses were recorded. The accuracy rates were compared using the Chi-squared test.

Results: The ChatGPT provided correct answers to 52.8% of the questions, while Bard answered 38.9% correctly, and Bing answered 44.4% correctly. All three AI programs provided identical responses to 20 (55.6%) of the questions, with 45% of these responses being correct and 55% incorrect. No significant difference was observed between the correct and incorrect responses from the three AI chatbots (P = 0.654).

Conclusion: AI chatbots should be developed to provide widespread access to accurate information about ocular inflammation, uveal diseases, and treatment modalities. Future research could explore ways to enhance the performance of these chatbots.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
研究人工智能程序在评估眼部炎症、葡萄膜疾病和治疗方法相关知识水平方面的比较优势。
目的:本研究的目的是评估 Chat Generative Pretrained Transformer(ChatGPT)、Bard 和 Bing 人工智能(AI)聊天机器人对眼部炎症、葡萄膜疾病和治疗方式的了解程度,并研究它们之间的相对性能比较:向 ChatGPT、Bard 和 Bing 人工智能聊天机器人提出了 36 个与眼部炎症、葡萄膜疾病和治疗方式有关的问题,并记录了正确和错误的回答。使用卡方检验比较了正确率:结果:ChatGPT 提供了 52.8% 的正确答案,Bard 回答了 38.9% 的正确答案,Bing 回答了 44.4% 的正确答案。所有三个人工智能程序都对 20 个问题(55.6%)做出了相同的回答,其中 45% 回答正确,55% 回答错误。三个人工智能聊天机器人的正确回答和错误回答之间没有明显差异(P = 0.654):结论:应开发人工智能聊天机器人,以广泛提供有关眼部炎症、葡萄膜疾病和治疗方法的准确信息。未来的研究可以探索提高这些聊天机器人性能的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
1.80
自引率
9.10%
发文量
68
审稿时长
19 weeks
期刊最新文献
Advancing glaucoma care with big data and artificial intelligence innovations. Application of artificial intelligence in glaucoma care: An updated review. Artificial intelligence and big data integration in anterior segment imaging for glaucoma. Big data and electronic health records for glaucoma research. Big data for imaging assessment in glaucoma.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1