Evaluation of the reliability and readability of answers given by chatbots to frequently asked questions about endophthalmitis: A cross-sectional study on chatbots.

IF 2.2 3区 医学 Q2 HEALTH CARE SCIENCES & SERVICES Health Informatics Journal Pub Date : 2024-10-01 DOI:10.1177/14604582241304679
Suleyman Demir
{"title":"Evaluation of the reliability and readability of answers given by chatbots to frequently asked questions about endophthalmitis: A cross-sectional study on chatbots.","authors":"Suleyman Demir","doi":"10.1177/14604582241304679","DOIUrl":null,"url":null,"abstract":"<p><p><b>Objective:</b> This study aimed to investigate the accuracy, reliability, and readability of A-Eye Consult, ChatGPT-4.0, Google Gemini and Copilot AI large language models (LLMs) in responding to patient questions about endophthalmitis. <b>Methods:</b> The LLMs' responses to 25 questions about endophthalmitis, frequently asked by patients, were evaluated by two ophthalmologists using a five-point Likert scale, with scores ranging from 1-5. The DISCERN scale assessed the reliability of the LLMs' responses, whereas the Flesch Reading Ease (FRE) and Flesch-Kincaid Grade Level (FKGL) indices assessed readability and text complexity, respectively. <b>Results:</b> A-Eye Consult and ChatGPT-4.0 outperformed Google Gemini and Copilot in providing comprehensive and precise responses. The Likert score significantly differed across all four LLMs (<i>p</i> < .001), with A-Eye Consult scoring significantly higher than Google Gemini and Copilot (<i>p</i> < .001). <b>Conclusions:</b> A-Eye Consult and ChatGPT-4.0 responses, while more complex than those of other LLMs, provided more reliable and accurate information.</p>","PeriodicalId":55069,"journal":{"name":"Health Informatics Journal","volume":"30 4","pages":"14604582241304679"},"PeriodicalIF":2.2000,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Health Informatics Journal","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1177/14604582241304679","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"HEALTH CARE SCIENCES & SERVICES","Score":null,"Total":0}
引用次数: 0

Abstract

Objective: This study aimed to investigate the accuracy, reliability, and readability of A-Eye Consult, ChatGPT-4.0, Google Gemini and Copilot AI large language models (LLMs) in responding to patient questions about endophthalmitis. Methods: The LLMs' responses to 25 questions about endophthalmitis, frequently asked by patients, were evaluated by two ophthalmologists using a five-point Likert scale, with scores ranging from 1-5. The DISCERN scale assessed the reliability of the LLMs' responses, whereas the Flesch Reading Ease (FRE) and Flesch-Kincaid Grade Level (FKGL) indices assessed readability and text complexity, respectively. Results: A-Eye Consult and ChatGPT-4.0 outperformed Google Gemini and Copilot in providing comprehensive and precise responses. The Likert score significantly differed across all four LLMs (p < .001), with A-Eye Consult scoring significantly higher than Google Gemini and Copilot (p < .001). Conclusions: A-Eye Consult and ChatGPT-4.0 responses, while more complex than those of other LLMs, provided more reliable and accurate information.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
评价聊天机器人对眼内炎常见问题的回答的可靠性和可读性:一项关于聊天机器人的横断面研究。
目的:本研究旨在探讨A-Eye Consult、ChatGPT-4.0、谷歌Gemini和Copilot AI大语言模型(llm)在回答患者关于眼内炎的问题时的准确性、可靠性和可读性。方法:两位眼科医生采用李克特五分制对LLMs对患者常问的关于眼内炎的25个问题的回答进行评估,评分范围为1-5分。辨别量表评估法学硕士回答的可靠性,而Flesch Reading Ease (FRE)和Flesch- kincaid Grade Level (FKGL)指数分别评估可读性和文本复杂性。结果:A-Eye Consult和ChatGPT-4.0在提供全面和精确的响应方面优于谷歌Gemini和Copilot。所有四种llm的Likert评分显著差异(p < 0.001), A-Eye Consult评分显著高于谷歌Gemini和Copilot (p < 0.001)。结论:A-Eye Consult和ChatGPT-4.0响应虽然比其他LLMs更为复杂,但提供的信息更为可靠和准确。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Health Informatics Journal
Health Informatics Journal HEALTH CARE SCIENCES & SERVICES-MEDICAL INFORMATICS
CiteScore
7.80
自引率
6.70%
发文量
80
审稿时长
6 months
期刊介绍: Health Informatics Journal is an international peer-reviewed journal. All papers submitted to Health Informatics Journal are subject to peer review by members of a carefully appointed editorial board. The journal operates a conventional single-blind reviewing policy in which the reviewer’s name is always concealed from the submitting author.
期刊最新文献
Researching public health datasets in the era of deep learning: a systematic literature review. A blueprint for large language model-augmented telehealth for HIV mitigation in Indonesia: A scoping review of a novel therapeutic modality. Advancing African American and hispanic health literacy with a bilingual, personalized, prevention smartphone application. Evaluating the quality of Spanish-language information for patients with type 2 diabetes on YouTube and Facebook. Pathways to usage intention of mobile health apps among hypertensive patients: A fuzzy-set qualitative comparative analysis.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1