American Academy of Orthopaedic Surgeons OrthoInfo provides more readable information regarding rotator cuff injury than ChatGPT

IF 3.3 Q1 ORTHOPEDICS Journal of ISAKOS Joint Disorders & Orthopaedic Sports Medicine Pub Date : 2025-06-01 Epub Date: 2025-02-12 DOI:10.1016/j.jisako.2025.100841
Catherine Hand , Camden Bohn , Shadia Tannir , Marisa Ulrich , Sami Saniei , Miguel Girod-Hoffman , Yining Lu , Brian Forsythe
{"title":"American Academy of Orthopaedic Surgeons OrthoInfo provides more readable information regarding rotator cuff injury than ChatGPT","authors":"Catherine Hand ,&nbsp;Camden Bohn ,&nbsp;Shadia Tannir ,&nbsp;Marisa Ulrich ,&nbsp;Sami Saniei ,&nbsp;Miguel Girod-Hoffman ,&nbsp;Yining Lu ,&nbsp;Brian Forsythe","doi":"10.1016/j.jisako.2025.100841","DOIUrl":null,"url":null,"abstract":"<div><h3>Introduction</h3><div>With over 61% of Americans seeking health information online, the accuracy and readability of this content are critical. Artificial intelligence (AI) tools, such as ChatGPT, have gained popularity in providing medical information, but concerns remain about their accessibility, especially for individuals with lower literacy levels. This study compares the readability and accuracy of ChatGPT-generated content with information from the American Academy of Orthopaedic Surgeons OrthoInfo website, focusing on rotator cuff injuries.</div></div><div><h3>Methods</h3><div>We formulated seven frequently asked questions about rotator cuff injuries, based on the OrthoInfo website, and gathered responses from both ChatGPT-4 and OrthoInfo. Readability was assessed using multiple readability metrics (Flesch-Kincaid, Gunning Fog, Coleman-Liau, Simple Measure of Gobbledygook Readability Formula, FORCAST Readability Formula, Fry Graph, and Raygor Readability Estimate), while accuracy was evaluated by three independent reviewers. Statistical analysis included <em>t</em>-tests and correlation analysis.</div></div><div><h3>Results</h3><div>ChatGPT responses required a higher education level to comprehend, with an average grade level of 14.7, compared to OrthoInfo's 11.9 (<em>p</em> ​&lt; ​0.01). The Flesch Reading Ease Index indicated that OrthoInfo's content (52.5) was more readable than ChatGPT's (25.9, <em>p</em> ​&lt; ​0.01). Both sources had high accuracy, with ChatGPT slightly lower in accuracy for the question about further damage to the rotator cuff (<em>p</em> ​&lt; ​0.05).</div></div><div><h3>Conclusion</h3><div>ChatGPT shows promise in delivering accurate health information but may not be suitable for all patients due to its higher complexity. A combination of AI and expert-reviewed, accessible content may enhance patient understanding and health literacy. Future developments should focus on improving AI's adaptability to different literacy levels.</div></div><div><h3>Level of evidence</h3><div>IV.</div></div>","PeriodicalId":36847,"journal":{"name":"Journal of ISAKOS Joint Disorders & Orthopaedic Sports Medicine","volume":"12 ","pages":"Article 100841"},"PeriodicalIF":3.3000,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of ISAKOS Joint Disorders & Orthopaedic Sports Medicine","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2059775425004584","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/12 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"ORTHOPEDICS","Score":null,"Total":0}
引用次数: 0

Abstract

Introduction

With over 61% of Americans seeking health information online, the accuracy and readability of this content are critical. Artificial intelligence (AI) tools, such as ChatGPT, have gained popularity in providing medical information, but concerns remain about their accessibility, especially for individuals with lower literacy levels. This study compares the readability and accuracy of ChatGPT-generated content with information from the American Academy of Orthopaedic Surgeons OrthoInfo website, focusing on rotator cuff injuries.

Methods

We formulated seven frequently asked questions about rotator cuff injuries, based on the OrthoInfo website, and gathered responses from both ChatGPT-4 and OrthoInfo. Readability was assessed using multiple readability metrics (Flesch-Kincaid, Gunning Fog, Coleman-Liau, Simple Measure of Gobbledygook Readability Formula, FORCAST Readability Formula, Fry Graph, and Raygor Readability Estimate), while accuracy was evaluated by three independent reviewers. Statistical analysis included t-tests and correlation analysis.

Results

ChatGPT responses required a higher education level to comprehend, with an average grade level of 14.7, compared to OrthoInfo's 11.9 (p ​< ​0.01). The Flesch Reading Ease Index indicated that OrthoInfo's content (52.5) was more readable than ChatGPT's (25.9, p ​< ​0.01). Both sources had high accuracy, with ChatGPT slightly lower in accuracy for the question about further damage to the rotator cuff (p ​< ​0.05).

Conclusion

ChatGPT shows promise in delivering accurate health information but may not be suitable for all patients due to its higher complexity. A combination of AI and expert-reviewed, accessible content may enhance patient understanding and health literacy. Future developments should focus on improving AI's adaptability to different literacy levels.

Level of evidence

IV.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
美国骨科学会OrthoInfo提供了比ChatGPT更可读的关于肩袖损伤的信息。
导读:超过61%的美国人在网上寻求健康信息,这些内容的准确性和可读性至关重要。像ChatGPT这样的人工智能工具在提供医疗信息方面越来越受欢迎,但人们仍然担心它们的可访问性,特别是对于文化水平较低的个人。本研究将chatgpt生成的内容的可读性和准确性与美国骨科学会(AAOS) OrthoInfo网站上的信息进行比较,重点关注肩袖损伤。方法:基于OrthoInfo网站,我们制定了关于肩袖损伤的7个常见问题,并收集了ChatGPT-4和OrthoInfo的回答。使用多种可读性指标(Flesch-Kincaid, Gunning Fog, Coleman-Liau, SMOG可读性公式,cast可读性公式,Fry Graph, Raygor可读性估计)评估可读性,而准确性由三位独立评审员评估。统计分析包括t检验和相关分析。结果:ChatGPT应答需要较高的教育程度才能理解,平均年级水平为14.7,而OrthoInfo的平均年级水平为11.9 (p < 0.01)。Flesch Reading Ease Index显示,OrthoInfo的内容(52.5)比ChatGPT的内容(25.9,p < 0.01)更具可读性。两个来源的准确性都很高,ChatGPT在关于肩袖进一步损伤的问题上的准确性略低(p < 0.05)。结论:ChatGPT有望提供准确的健康信息,但由于其较高的复杂性,可能不适合所有患者。人工智能与专家审查的可访问内容相结合,可以增强患者的理解和健康素养。未来的发展应侧重于提高人工智能对不同文化水平的适应能力。证据等级:四级。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
2.90
自引率
6.20%
发文量
61
审稿时长
108 days
期刊最新文献
Failed shoulder instability surgery: State -of-the-art Early arthroscopic fixation of low-grade juvenile osteochondritis dissecans with bioabsorbable pins achieves excellent clinical and imaging outcomes in high-demand athletes Equivocal physical activity outcomes 10 Years after patellofemoral vs total knee replacement: Follow-up of a previously reported randomized controlled trial in isolated patellofemoral joint osteoarthritis Development and validation of a scale to assess risk-taking intentions during return to sport following anterior cruciate ligament reconstruction Arthroscopic intercruciate trans-septal all-inside repair for lateral bucket-handle meniscus tear: A current technique
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1