基于 ChatGPT 的医学摘要总结的质量、准确性和偏差。

IF 4.4 2区 医学 Q1 MEDICINE, GENERAL & INTERNAL Annals of Family Medicine Pub Date : 2024-03-01 DOI:10.1370/afm.3075
Joel Hake, Miles Crowley, Allison Coy, Denton Shanks, Aundria Eoff, Kalee Kirmer-Voss, Gurpreet Dhanda, Daniel J Parente
{"title":"基于 ChatGPT 的医学摘要总结的质量、准确性和偏差。","authors":"Joel Hake, Miles Crowley, Allison Coy, Denton Shanks, Aundria Eoff, Kalee Kirmer-Voss, Gurpreet Dhanda, Daniel J Parente","doi":"10.1370/afm.3075","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>Worldwide clinical knowledge is expanding rapidly, but physicians have sparse time to review scientific literature. Large language models (eg, Chat Generative Pretrained Transformer [ChatGPT]), might help summarize and prioritize research articles to review. However, large language models sometimes \"hallucinate\" incorrect information.</p><p><strong>Methods: </strong>We evaluated ChatGPT's ability to summarize 140 peer-reviewed abstracts from 14 journals. Physicians rated the quality, accuracy, and bias of the ChatGPT summaries. We also compared human ratings of relevance to various areas of medicine to ChatGPT relevance ratings.</p><p><strong>Results: </strong>ChatGPT produced summaries that were 70% shorter (mean abstract length of 2,438 characters decreased to 739 characters). Summaries were nevertheless rated as high quality (median score 90, interquartile range [IQR] 87.0-92.5; scale 0-100), high accuracy (median 92.5, IQR 89.0-95.0), and low bias (median 0, IQR 0-7.5). Serious inaccuracies and hallucinations were uncommon. Classification of the relevance of entire journals to various fields of medicine closely mirrored physician classifications (nonlinear standard error of the regression [SER] 8.6 on a scale of 0-100). However, relevance classification for individual articles was much more modest (SER 22.3).</p><p><strong>Conclusions: </strong>Summaries generated by ChatGPT were 70% shorter than mean abstract length and were characterized by high quality, high accuracy, and low bias. Conversely, ChatGPT had modest ability to classify the relevance of articles to medical specialties. We suggest that ChatGPT can help family physicians accelerate review of the scientific literature and have developed software (pyJournalWatch) to support this application. Life-critical medical decisions should remain based on full, critical, and thoughtful evaluation of the full text of research articles in context with clinical guidelines.</p>","PeriodicalId":50973,"journal":{"name":"Annals of Family Medicine","volume":null,"pages":null},"PeriodicalIF":4.4000,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11237196/pdf/","citationCount":"0","resultStr":"{\"title\":\"Quality, Accuracy, and Bias in ChatGPT-Based Summarization of Medical Abstracts.\",\"authors\":\"Joel Hake, Miles Crowley, Allison Coy, Denton Shanks, Aundria Eoff, Kalee Kirmer-Voss, Gurpreet Dhanda, Daniel J Parente\",\"doi\":\"10.1370/afm.3075\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Purpose: </strong>Worldwide clinical knowledge is expanding rapidly, but physicians have sparse time to review scientific literature. Large language models (eg, Chat Generative Pretrained Transformer [ChatGPT]), might help summarize and prioritize research articles to review. However, large language models sometimes \\\"hallucinate\\\" incorrect information.</p><p><strong>Methods: </strong>We evaluated ChatGPT's ability to summarize 140 peer-reviewed abstracts from 14 journals. Physicians rated the quality, accuracy, and bias of the ChatGPT summaries. We also compared human ratings of relevance to various areas of medicine to ChatGPT relevance ratings.</p><p><strong>Results: </strong>ChatGPT produced summaries that were 70% shorter (mean abstract length of 2,438 characters decreased to 739 characters). Summaries were nevertheless rated as high quality (median score 90, interquartile range [IQR] 87.0-92.5; scale 0-100), high accuracy (median 92.5, IQR 89.0-95.0), and low bias (median 0, IQR 0-7.5). Serious inaccuracies and hallucinations were uncommon. Classification of the relevance of entire journals to various fields of medicine closely mirrored physician classifications (nonlinear standard error of the regression [SER] 8.6 on a scale of 0-100). However, relevance classification for individual articles was much more modest (SER 22.3).</p><p><strong>Conclusions: </strong>Summaries generated by ChatGPT were 70% shorter than mean abstract length and were characterized by high quality, high accuracy, and low bias. Conversely, ChatGPT had modest ability to classify the relevance of articles to medical specialties. We suggest that ChatGPT can help family physicians accelerate review of the scientific literature and have developed software (pyJournalWatch) to support this application. Life-critical medical decisions should remain based on full, critical, and thoughtful evaluation of the full text of research articles in context with clinical guidelines.</p>\",\"PeriodicalId\":50973,\"journal\":{\"name\":\"Annals of Family Medicine\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":4.4000,\"publicationDate\":\"2024-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11237196/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Annals of Family Medicine\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1370/afm.3075\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MEDICINE, GENERAL & INTERNAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Annals of Family Medicine","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1370/afm.3075","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MEDICINE, GENERAL & INTERNAL","Score":null,"Total":0}
引用次数: 0

摘要

目的:世界范围内的临床知识正在迅速扩展,但医生用于审阅科学文献的时间却非常有限。大型语言模型(如 Chat Generative Pretrained Transformer [ChatGPT])可以帮助总结和优先排序需要审阅的研究文章。然而,大型语言模型有时会 "幻化 "出不正确的信息:我们评估了 ChatGPT 总结 14 种期刊 140 篇同行评审摘要的能力。医生对 ChatGPT 摘要的质量、准确性和偏差进行了评分。我们还将人类对不同医学领域的相关性评级与 ChatGPT 的相关性评级进行了比较:结果:ChatGPT 生成的摘要缩短了 70%(摘要平均长度从 2438 个字符减少到 739 个字符)。尽管如此,摘要仍被评为高质量(中位数 90 分,四分位数间距 [IQR] 87.0-92.5;评分标准 0-100)、高准确性(中位数 92.5 分,IQR 89.0-95.0)和低偏差(中位数 0 分,IQR 0-7.5)。严重误差和幻觉并不常见。整本期刊与不同医学领域的相关性分类与医生的分类非常接近(非线性回归标准误差 [SER] 8.6,0-100 分)。然而,单篇文章的相关性分类则要适中得多(SER 22.3):结论:ChatGPT 生成的摘要比平均摘要长度短 70%,而且具有高质量、高准确性和低偏差的特点。相反,ChatGPT 对文章与医学专业相关性的分类能力一般。我们建议 ChatGPT 可以帮助家庭医生加快科学文献的审阅速度,并已开发出支持该应用的软件(pyJournalWatch)。生命攸关的医疗决策仍应建立在根据临床指南对研究文章全文进行全面、批判性和深思熟虑的评估基础之上。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Quality, Accuracy, and Bias in ChatGPT-Based Summarization of Medical Abstracts.

Purpose: Worldwide clinical knowledge is expanding rapidly, but physicians have sparse time to review scientific literature. Large language models (eg, Chat Generative Pretrained Transformer [ChatGPT]), might help summarize and prioritize research articles to review. However, large language models sometimes "hallucinate" incorrect information.

Methods: We evaluated ChatGPT's ability to summarize 140 peer-reviewed abstracts from 14 journals. Physicians rated the quality, accuracy, and bias of the ChatGPT summaries. We also compared human ratings of relevance to various areas of medicine to ChatGPT relevance ratings.

Results: ChatGPT produced summaries that were 70% shorter (mean abstract length of 2,438 characters decreased to 739 characters). Summaries were nevertheless rated as high quality (median score 90, interquartile range [IQR] 87.0-92.5; scale 0-100), high accuracy (median 92.5, IQR 89.0-95.0), and low bias (median 0, IQR 0-7.5). Serious inaccuracies and hallucinations were uncommon. Classification of the relevance of entire journals to various fields of medicine closely mirrored physician classifications (nonlinear standard error of the regression [SER] 8.6 on a scale of 0-100). However, relevance classification for individual articles was much more modest (SER 22.3).

Conclusions: Summaries generated by ChatGPT were 70% shorter than mean abstract length and were characterized by high quality, high accuracy, and low bias. Conversely, ChatGPT had modest ability to classify the relevance of articles to medical specialties. We suggest that ChatGPT can help family physicians accelerate review of the scientific literature and have developed software (pyJournalWatch) to support this application. Life-critical medical decisions should remain based on full, critical, and thoughtful evaluation of the full text of research articles in context with clinical guidelines.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Annals of Family Medicine
Annals of Family Medicine 医学-医学:内科
CiteScore
3.70
自引率
4.50%
发文量
142
审稿时长
6-12 weeks
期刊介绍: The Annals of Family Medicine is a peer-reviewed research journal to meet the needs of scientists, practitioners, policymakers, and the patients and communities they serve.
期刊最新文献
Prehospital Clinical Decision-Making for Medication Administration for Behavioral Emergencies. Rapid Antigen Testing in School Health Offices for Surveillance of SARS-CoV-2 and Influenza A. Assessment and comparison of the ILI case definition in clinical and school-based community settings: ORCHARDS/IISP. Efficacy and Safety of BB-12 Supplemented Strawberry Yogurt For Healthy Children on Antibiotics (PLAY ON). Exploratory Pilot Studies to Demonstrate Mechanisms of Preventing Antibiotic-Associated Diarrhea and the Role for Probiotics.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1