Testing ChatGPT's Ability to Provide Patient and Physician Information on Aortic Aneurysm

IF 1.7 3区 医学 Q2 SURGERY Journal of Surgical Research Pub Date : 2025-03-01 Epub Date: 2025-02-27 DOI:10.1016/j.jss.2025.01.015
Daniel J. Bertges MD , Adam W. Beck MD , Marc Schermerhorn MD , Mark K. Eskandari MD , Jens Eldrup-Jorgensen MD , Sean Liebscher MD , Robyn Guinto MD , Mead Ferris MD , Andy Stanley MD , Georg Steinthorsson MD , Matthew Alef MD , Salvatore T. Scali MD
{"title":"Testing ChatGPT's Ability to Provide Patient and Physician Information on Aortic Aneurysm","authors":"Daniel J. Bertges MD ,&nbsp;Adam W. Beck MD ,&nbsp;Marc Schermerhorn MD ,&nbsp;Mark K. Eskandari MD ,&nbsp;Jens Eldrup-Jorgensen MD ,&nbsp;Sean Liebscher MD ,&nbsp;Robyn Guinto MD ,&nbsp;Mead Ferris MD ,&nbsp;Andy Stanley MD ,&nbsp;Georg Steinthorsson MD ,&nbsp;Matthew Alef MD ,&nbsp;Salvatore T. Scali MD","doi":"10.1016/j.jss.2025.01.015","DOIUrl":null,"url":null,"abstract":"<div><h3>Introduction</h3><div>Our objective was to test the ability of ChatGPT 4.0 to provide accurate information for patients and physicians about abdominal aortic aneurysms (AAA) and to assess its alignment with Society for Vascular Surgery (SVS) clinical practice guidelines (CPG) for AAA care.</div></div><div><h3>Material and methods</h3><div>Fifteen patient-level questions, 37 questions selected to reflect 28 SVS CPGs and 4 questions regarding AAA rupture risk were posed to ChatGPT 4.0. Single responses were recorded and graded for accuracy and quality by ten board-certified vascular surgeons as well as two fellow trainees using a 5-point Likert scale; 1 = very poor, 2 = poor, 3 = fair, 4 = good, and 5 = excellent.</div></div><div><h3>Results</h3><div>The mean of the means (MoM) accuracy rating across all 15 patient-level questions was 4.4 (SD 0.4, quartile range (QR) 4.2-4.7). ChatGPT 4.0 demonstrated good alignment with SVS practice guidelines (MoM: 4.2, SD: 0.4, QR: 3.9-4.5). The accuracy of responses was consistent across guideline categories; screening or surveillance (4.2), indications for surgery (4.5), preoperative risk assessment (4.5), perioperative coronary revascularization (4.1), and perioperative management (4.2). The generative artificial intelligence bot demonstrated only fair performance in answering the annual AAA rupture risk (MoM: 3.4, SD: 1.2, QR: 2.3-4.3).</div></div><div><h3>Conclusions</h3><div>ChatGPT 4.0 provided accurate responses to a variety of patient-level questions regarding AAA. Responses were well-aligned with current SVS CPGs except for inaccuracies in the risk of AAA rupture at varying diameters. The emergence of generative artificial intelligence bots presents an opportunity for study of applications in patient education and to determine their ability to augment the vascular specialist's knowledge base.</div></div>","PeriodicalId":17030,"journal":{"name":"Journal of Surgical Research","volume":"307 ","pages":"Pages 129-138"},"PeriodicalIF":1.7000,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Surgical Research","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0022480425000332","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/27 0:00:00","PubModel":"Epub","JCR":"Q2","JCRName":"SURGERY","Score":null,"Total":0}
引用次数: 0

Abstract

Introduction

Our objective was to test the ability of ChatGPT 4.0 to provide accurate information for patients and physicians about abdominal aortic aneurysms (AAA) and to assess its alignment with Society for Vascular Surgery (SVS) clinical practice guidelines (CPG) for AAA care.

Material and methods

Fifteen patient-level questions, 37 questions selected to reflect 28 SVS CPGs and 4 questions regarding AAA rupture risk were posed to ChatGPT 4.0. Single responses were recorded and graded for accuracy and quality by ten board-certified vascular surgeons as well as two fellow trainees using a 5-point Likert scale; 1 = very poor, 2 = poor, 3 = fair, 4 = good, and 5 = excellent.

Results

The mean of the means (MoM) accuracy rating across all 15 patient-level questions was 4.4 (SD 0.4, quartile range (QR) 4.2-4.7). ChatGPT 4.0 demonstrated good alignment with SVS practice guidelines (MoM: 4.2, SD: 0.4, QR: 3.9-4.5). The accuracy of responses was consistent across guideline categories; screening or surveillance (4.2), indications for surgery (4.5), preoperative risk assessment (4.5), perioperative coronary revascularization (4.1), and perioperative management (4.2). The generative artificial intelligence bot demonstrated only fair performance in answering the annual AAA rupture risk (MoM: 3.4, SD: 1.2, QR: 2.3-4.3).

Conclusions

ChatGPT 4.0 provided accurate responses to a variety of patient-level questions regarding AAA. Responses were well-aligned with current SVS CPGs except for inaccuracies in the risk of AAA rupture at varying diameters. The emergence of generative artificial intelligence bots presents an opportunity for study of applications in patient education and to determine their ability to augment the vascular specialist's knowledge base.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
测试ChatGPT为患者和医生提供主动脉瘤信息的能力
我们的目的是测试ChatGPT 4.0为腹主动脉瘤(AAA)患者和医生提供准确信息的能力,并评估其与血管外科学会(SVS)临床实践指南(CPG)对AAA护理的一致性。材料和方法在ChatGPT 4.0中提出15个患者层面的问题、37个反映28个SVS CPGs的问题和4个关于AAA破裂风险的问题。单次回答由10名委员会认证的血管外科医生和两名学员使用5分李克特量表进行记录和准确性和质量评分;1 =非常贫穷,2 =差,3 =公平,4 =好,5 =优秀。结果15个患者层面问题的平均(MoM)准确率为4.4 (SD 0.4,四分位数范围4.2 ~ 4.7)。ChatGPT 4.0表现出与SVS实践指南的良好一致性(MoM: 4.2, SD: 0.4, QR: 3.9-4.5)。回答的准确性在不同的指南类别中是一致的;筛查或监测(4.2),手术指征(4.5),术前风险评估(4.5),围手术期冠状动脉重建术(4.1)和围手术期管理(4.2)。生成式人工智能机器人在回答年度AAA破裂风险(MoM: 3.4, SD: 1.2, QR: 2.3-4.3)方面表现一般。结论:schatgpt 4.0对AAA患者层面的各种问题提供了准确的答案,除了不同直径AAA破裂风险的不准确性外,答案与目前的SVS CPGs一致。生成式人工智能机器人的出现为研究患者教育中的应用提供了机会,并确定了它们增强血管专家知识库的能力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
3.90
自引率
4.50%
发文量
627
审稿时长
138 days
期刊介绍: The Journal of Surgical Research: Clinical and Laboratory Investigation publishes original articles concerned with clinical and laboratory investigations relevant to surgical practice and teaching. The journal emphasizes reports of clinical investigations or fundamental research bearing directly on surgical management that will be of general interest to a broad range of surgeons and surgical researchers. The articles presented need not have been the products of surgeons or of surgical laboratories. The Journal of Surgical Research also features review articles and special articles relating to educational, research, or social issues of interest to the academic surgical community.
期刊最新文献
ReEnSta Alleviates Pain by Reducing Postoperative Swelling and Blood Stasis After Open Surgery Mechanisms of Hydrogel Vascularization Lymphedema Education Advocacy Program (LEAP): Addressing Gaps in Patient Education of Breast Cancer-Related Lymphedema (BCRL) Video-Assisted Lateral Neck Dissection has Good Efficacy and Safety: A Long-Term Follow-Up Study Neoadjuvant Radiation is Causally Linked to Increased Operative Time and Perioperative Blood Transfusion in Pancreatic Ductal Adenocarcinoma
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1