Evaluating ChatGPT to test its robustness as an interactive information database of radiation oncology and to assess its responses to common queries from radiotherapy patients: A single institution investigation

IF 1.5 4区 医学 Q4 ONCOLOGY Cancer Radiotherapie Pub Date : 2024-06-01 DOI:10.1016/j.canrad.2023.11.005
V.K. Pandey , A. Munshi , B.K. Mohanti , K. Bansal , K. Rastogi
{"title":"Evaluating ChatGPT to test its robustness as an interactive information database of radiation oncology and to assess its responses to common queries from radiotherapy patients: A single institution investigation","authors":"V.K. Pandey ,&nbsp;A. Munshi ,&nbsp;B.K. Mohanti ,&nbsp;K. Bansal ,&nbsp;K. Rastogi","doi":"10.1016/j.canrad.2023.11.005","DOIUrl":null,"url":null,"abstract":"<div><h3>Purpose</h3><p>Commercial vendors have created artificial intelligence (AI) tools for use in all aspects of life and medicine, including radiation oncology. AI innovations will likely disrupt workflows in the field of radiation oncology. However, limited data exist on using AI-based chatbots about the quality of radiation oncology information. This study aims to assess the accuracy of ChatGPT, an AI-based chatbot, in answering patients’ questions during their first visit to the radiation oncology outpatient department and test knowledge of ChatGPT in radiation oncology.</p></div><div><h3>Material and methods</h3><p>Expert opinion was formulated using a set of ten standard questions of patients encountered in outpatient department practice. A blinded expert opinion was taken for the ten questions on common queries of patients in outpatient department visits, and the same questions were evaluated on ChatGPT version 3.5 (ChatGPT 3.5). The answers by expert and ChatGPT were independently evaluated for accuracy by three scientific reviewers. Additionally, a comparison was made for the extent of similarity of answers between ChatGPT and experts by a response scoring for each answer. Word count and Flesch-Kincaid readability score and grade were done for the responses obtained from expert and ChatGPT. A comparison of the answers of ChatGPT and expert was done with a Likert scale. As a second component of the study, we tested the technical knowledge of ChatGPT. Ten multiple choice questions were framed with increasing order of difficulty – basic, intermediate and advanced, and the responses were evaluated on ChatGPT. Statistical testing was done using SPSS version 27.</p></div><div><h3>Results</h3><p>After expert review, the accuracy of expert opinion was 100%, and ChatGPT's was 80% (8/10) for regular questions encountered in outpatient department visits. A noticeable difference was observed in word count and readability of answers from expert opinion or ChatGPT. Of the ten multiple-choice questions for assessment of radiation oncology database, ChatGPT had an accuracy rate of 90% (9 out of 10). One answer to a basic-level question was incorrect, whereas all answers to intermediate and difficult-level questions were correct.</p></div><div><h3>Conclusion</h3><p>ChatGPT provides reasonably accurate information about routine questions encountered in the first outpatient department visit of the patient and also demonstrated a sound knowledge of the subject. The result of our study can inform the future development of educational tools in radiation oncology and may have implications in other medical fields. This is the first study that provides essential insight into the potentially positive capabilities of two components of ChatGPT: firstly, ChatGPT's response to common queries of patients at OPD visits, and secondly, the assessment of the radiation oncology knowledge base of ChatGPT.</p></div>","PeriodicalId":9504,"journal":{"name":"Cancer Radiotherapie","volume":"28 3","pages":"Pages 258-264"},"PeriodicalIF":1.5000,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Cancer Radiotherapie","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1278321824000611","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"ONCOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose

Commercial vendors have created artificial intelligence (AI) tools for use in all aspects of life and medicine, including radiation oncology. AI innovations will likely disrupt workflows in the field of radiation oncology. However, limited data exist on using AI-based chatbots about the quality of radiation oncology information. This study aims to assess the accuracy of ChatGPT, an AI-based chatbot, in answering patients’ questions during their first visit to the radiation oncology outpatient department and test knowledge of ChatGPT in radiation oncology.

Material and methods

Expert opinion was formulated using a set of ten standard questions of patients encountered in outpatient department practice. A blinded expert opinion was taken for the ten questions on common queries of patients in outpatient department visits, and the same questions were evaluated on ChatGPT version 3.5 (ChatGPT 3.5). The answers by expert and ChatGPT were independently evaluated for accuracy by three scientific reviewers. Additionally, a comparison was made for the extent of similarity of answers between ChatGPT and experts by a response scoring for each answer. Word count and Flesch-Kincaid readability score and grade were done for the responses obtained from expert and ChatGPT. A comparison of the answers of ChatGPT and expert was done with a Likert scale. As a second component of the study, we tested the technical knowledge of ChatGPT. Ten multiple choice questions were framed with increasing order of difficulty – basic, intermediate and advanced, and the responses were evaluated on ChatGPT. Statistical testing was done using SPSS version 27.

Results

After expert review, the accuracy of expert opinion was 100%, and ChatGPT's was 80% (8/10) for regular questions encountered in outpatient department visits. A noticeable difference was observed in word count and readability of answers from expert opinion or ChatGPT. Of the ten multiple-choice questions for assessment of radiation oncology database, ChatGPT had an accuracy rate of 90% (9 out of 10). One answer to a basic-level question was incorrect, whereas all answers to intermediate and difficult-level questions were correct.

Conclusion

ChatGPT provides reasonably accurate information about routine questions encountered in the first outpatient department visit of the patient and also demonstrated a sound knowledge of the subject. The result of our study can inform the future development of educational tools in radiation oncology and may have implications in other medical fields. This is the first study that provides essential insight into the potentially positive capabilities of two components of ChatGPT: firstly, ChatGPT's response to common queries of patients at OPD visits, and secondly, the assessment of the radiation oncology knowledge base of ChatGPT.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
对 ChatGPT 进行评估,以测试其作为放射肿瘤学互动信息数据库的稳健性,并评估其对放射治疗患者常见询问的回应:单一机构调查。
目的:商业供应商已开发出人工智能(AI)工具,可用于生活和医学的各个方面,包括肿瘤放射学。人工智能的创新很可能会颠覆放射肿瘤学领域的工作流程。然而,关于使用基于人工智能的聊天机器人获取放射肿瘤学信息质量的数据还很有限。本研究旨在评估基于人工智能的聊天机器人 ChatGPT 在回答首次到放射肿瘤学门诊部就诊的患者问题时的准确性,并测试 ChatGPT 在放射肿瘤学方面的知识:使用门诊部实践中遇到的患者的十个标准问题,形成专家意见。就门诊部就诊患者的常见疑问对这十个问题进行了盲法专家鉴定,并在 ChatGPT 3.5 版(ChatGPT 3.5)上对相同的问题进行了评估。专家和 ChatGPT 的答案由三位科学评审员独立评估其准确性。此外,ChatGPT 和专家答案的相似程度也通过对每个答案的评分进行了比较。对专家和 ChatGPT 的回答进行了字数统计、Flesch-Kincaid 可读性评分和等级评定。通过李克特量表对 ChatGPT 和专家的答案进行了比较。作为研究的第二部分,我们测试了 ChatGPT 的技术知识。我们设计了十道选择题,难度依次为基础、中级和高级,并对 ChatGPT 的回答进行了评估。统计测试使用 SPSS 27 版本:经过专家评审,对于门诊部就诊中遇到的常规问题,专家意见的准确率为 100%,而 ChatGPT 的准确率为 80%(8/10)。专家意见和 ChatGPT 的答案在字数和可读性方面存在明显差异。在放射肿瘤学数据库评估的 10 道多选题中,ChatGPT 的准确率为 90%(10 选 9)。基础题有一个答案是错误的,而中级和困难题的所有答案都是正确的:结论:ChatGPT 为患者在首次门诊就诊时遇到的常规问题提供了相当准确的信息,同时也展示了对该主题的充分了解。我们的研究结果可为今后放射肿瘤学教育工具的开发提供参考,并可能对其他医学领域产生影响。这是第一项对 ChatGPT 两个组成部分的潜在积极功能提供重要见解的研究:首先是 ChatGPT 对患者在门诊就诊时常见问题的回答,其次是对 ChatGPT 放射肿瘤学知识库的评估。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Cancer Radiotherapie
Cancer Radiotherapie 医学-核医学
CiteScore
2.20
自引率
23.10%
发文量
129
审稿时长
63 days
期刊介绍: Cancer/radiothérapie se veut d''abord et avant tout un organe francophone de publication des travaux de recherche en radiothérapie. La revue a pour objectif de diffuser les informations majeures sur les travaux de recherche en cancérologie et tout ce qui touche de près ou de loin au traitement du cancer par les radiations : technologie, radiophysique, radiobiologie et radiothérapie clinique.
期刊最新文献
Long-term cardiac mortality in patients treated with radiation for gastric mucosa-associated lymphoid tissue lymphoma Dose comparison between hybrid volumetric modulated arc therapy, volumetric modulated arc therapy, and three-dimensional conformal radiotherapy for breast/chest wall irradiation, including regional lymph node irradiation using deep inspiration breath-hold technique Cyberneo trial investigating the efficacy of stereotactic radiotherapy combined to neoadjuvant chemotherapy for locally advanced breast cancer: 14-years follow-up results Exploring treatment complexity in maxillary ameloblastoma: A case study on the efficacy of radiotherapy with transcriptomic and literature insights Does prior radiotherapy impact the acute cellular liver graft rejection?
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1