Translating ophthalmic medical jargon with artificial intelligence: a comparative comprehension study.

Michael Balas, Alexander J Kaplan, Kaisra Esmail, Solin Saleh, Rahul A Sharma, Peng Yan, Parnian Arjmand
{"title":"Translating ophthalmic medical jargon with artificial intelligence: a comparative comprehension study.","authors":"Michael Balas, Alexander J Kaplan, Kaisra Esmail, Solin Saleh, Rahul A Sharma, Peng Yan, Parnian Arjmand","doi":"10.1016/j.jcjo.2024.11.003","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>Our goal was to evaluate the efficacy of OpenAI's ChatGPT-4.0 large language model (LLM) in translating technical ophthalmology terminology into more comprehensible language for allied health care professionals and compare it with other LLMs.</p><p><strong>Design: </strong>Observational cross-sectional study.</p><p><strong>Participants: </strong>Five ophthalmologists each contributed three clinical encounter notes, totaling 15 reports for analysis.</p><p><strong>Methods: </strong>Notes were translated into more comprehensible language using ChatGPT-4.0, ChatGPT-4o, Claude 3 Sonnet, and Google Gemini. Ten family physicians, masked to whether the note was original or translated by ChatGPT-4.0, independently evaluated both sets using Likert scales to assess comprehension and utility for clinical decision-making. Readability was evaluated using Flesch Reading Ease and Flesch-Kincaid Grade Level scores. Five ophthalmologist raters compared performance between LLMs and identified translation errors.</p><p><strong>Results: </strong>ChatGPT-4.0 translations significantly outperformed the original notes in terms of comprehension (mean score of 4.7/5.0 vs 3.7/5.0; p < 0.001) and perceived usefulness (mean score of 4.6/5.0 vs 3.8/5.0; p < 0.005). Readability analysis demonstrated mildly increased linguistic complexity in the translated notes. ChatGPT-4.0 was preferred in 8 of 15 cases, ChatGPT-4o in 4, Gemini in 3, and Claude 3 Sonnet in 0 cases. All models exhibited some translation errors, but ChatGPT-4o and ChatGPT-4.0 had fewer inaccuracies.</p><p><strong>Conclusions: </strong>ChatGPT-4.0 can significantly enhance the comprehensibility of ophthalmic notes, facilitating better interprofessional communication and suggesting a promising role for LLMs in medical translation. However, the results also underscore the need for ongoing refinement and careful implementation of such technologies. Further research is needed to validate these findings across a broader range of specialties and languages.</p>","PeriodicalId":9606,"journal":{"name":"Canadian journal of ophthalmology. Journal canadien d'ophtalmologie","volume":" ","pages":""},"PeriodicalIF":3.3000,"publicationDate":"2024-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Canadian journal of ophthalmology. Journal canadien d'ophtalmologie","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1016/j.jcjo.2024.11.003","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Objective: Our goal was to evaluate the efficacy of OpenAI's ChatGPT-4.0 large language model (LLM) in translating technical ophthalmology terminology into more comprehensible language for allied health care professionals and compare it with other LLMs.

Design: Observational cross-sectional study.

Participants: Five ophthalmologists each contributed three clinical encounter notes, totaling 15 reports for analysis.

Methods: Notes were translated into more comprehensible language using ChatGPT-4.0, ChatGPT-4o, Claude 3 Sonnet, and Google Gemini. Ten family physicians, masked to whether the note was original or translated by ChatGPT-4.0, independently evaluated both sets using Likert scales to assess comprehension and utility for clinical decision-making. Readability was evaluated using Flesch Reading Ease and Flesch-Kincaid Grade Level scores. Five ophthalmologist raters compared performance between LLMs and identified translation errors.

Results: ChatGPT-4.0 translations significantly outperformed the original notes in terms of comprehension (mean score of 4.7/5.0 vs 3.7/5.0; p < 0.001) and perceived usefulness (mean score of 4.6/5.0 vs 3.8/5.0; p < 0.005). Readability analysis demonstrated mildly increased linguistic complexity in the translated notes. ChatGPT-4.0 was preferred in 8 of 15 cases, ChatGPT-4o in 4, Gemini in 3, and Claude 3 Sonnet in 0 cases. All models exhibited some translation errors, but ChatGPT-4o and ChatGPT-4.0 had fewer inaccuracies.

Conclusions: ChatGPT-4.0 can significantly enhance the comprehensibility of ophthalmic notes, facilitating better interprofessional communication and suggesting a promising role for LLMs in medical translation. However, the results also underscore the need for ongoing refinement and careful implementation of such technologies. Further research is needed to validate these findings across a broader range of specialties and languages.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
目标:我们的目标是评估 OpenAI 的 ChatGPT-4.0 大语言模型(LLM)在将眼科专业术语翻译成专职医疗人员更易理解的语言方面的功效,并将其与其他 LLM 进行比较:观察性横断面研究:五名眼科医生每人提供三份临床诊疗记录,共 15 份报告供分析:使用 ChatGPT-4.0、ChatGPT-4o、Claude 3 Sonnet 和 Google Gemini 将笔记翻译成更易理解的语言。十位家庭医生在不知道笔记是原文还是由 ChatGPT-4.0 翻译的情况下,使用李克特量表对两套笔记进行独立评估,以评估其理解能力和对临床决策的实用性。可读性采用 Flesch 阅读轻松度和 Flesch-Kincaid 等级评分进行评估。五位眼科医生评分员比较了 LLM 之间的表现,并找出了翻译错误:结果:ChatGPT-4.0 译文在理解能力(平均分 4.7/5.0 vs 3.7/5.0;p < 0.001)和感知有用性(平均分 4.6/5.0 vs 3.8/5.0;p < 0.005)方面明显优于原注释。可读性分析表明,翻译笔记的语言复杂性略有增加。在 15 个案例中,有 8 个案例首选 ChatGPT-4.0,4 个案例首选 ChatGPT-4o,3 个案例首选 Gemini,0 个案例首选 Claude 3 Sonnet。所有模型都出现了一些翻译错误,但 ChatGPT-4o 和 ChatGPT-4.0 的错误率较低:结论:ChatGPT-4.0 能显著提高眼科笔记的可理解性,促进更好的跨专业交流,并表明语言学硕士在医学翻译中大有可为。然而,研究结果也强调了对此类技术进行不断改进和谨慎实施的必要性。要在更广泛的专业和语言范围内验证这些研究结果,还需要进一步的研究。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
3.20
自引率
4.80%
发文量
223
审稿时长
38 days
期刊介绍: Official journal of the Canadian Ophthalmological Society. The Canadian Journal of Ophthalmology (CJO) is the official journal of the Canadian Ophthalmological Society and is committed to timely publication of original, peer-reviewed ophthalmology and vision science articles.
期刊最新文献
Unequal access to routine eye exams in Canada: an analysis of government-funded eye exam coverage policy and associated vision health outcomes. Crafting the ophthalmic surgeon's mindset. Inpatient topical glaucoma medication-ordering discrepancies: a study of frequency and risk factors for inaccurate reconciliation. Spontaneous Resolution of Eyelid Diffuse Large B-Cell Lymphoma after Incisional Biopsy. Sympathetic ophthalmia in a patient with ciliary body melanoma and extraocular extension.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1