{"title":"评估在一个流行的、可公开访问的人工智能模型(GPT-4)上获得的有关心血管疾病预防的错误信息","authors":"","doi":"10.1016/j.ajpc.2024.100806","DOIUrl":null,"url":null,"abstract":"<div><h3>Therapeutic Area</h3><div>Other: Artificial intelligence; Misinformation</div></div><div><h3>Background</h3><div>Misinformation regarding CVD prevention is prevalent on the internet and on social media. Chat-based artificial intelligence (AI) models such as ChatGPT have gained over 100 million users, are publicly accessible, and may provide appropriate information for simple CVD prevention topics. Whether these public AI models may propagate misinformation regarding CVD prevention is uncertain.</div></div><div><h3>Methods</h3><div>This study was performed in March 2024 using the subscription-based version of GPT-4 (OpenAI, USA). Prompts regarding six CVD prevention topics (statin therapy and muscle-side effects, dementia, and liver disease; fish oil; supplements; and low-density lipoprotein-cholesterol and heart disease) were posed. Prompts were framed in two tones: a neutral tone and a misinformation-prompting tone. The misinformation-prompting tone requested specific arguments and scientific references to support misinformation. Each tone and topic was prompted in a different chatbot instance. Each response was reviewed by a board-certified cardiologist specializing in preventive cardiology at a tertiary care center. If a response had multiple bullet-points with individual scientific references, each bullet-point was graded separately. Responses were graded as appropriate (accurate content and references), borderline (minor inaccuracies or references published >20 years ago), or inappropriate (inaccurate content and/or references, including non-existent references).</div></div><div><h3>Results</h3><div>For the six prompts posed with a neutral tone, all responses lacked scientific references and were graded as appropriate (100%). For all six prompts posed with a misinformation-prompting tone, each response consisted of multiple discrete bullet-points with a scientific reference for each individual point. Of 31 bullet-points across the six topics obtained using a misinformation-prompting tone, 32.2% (10/31) were graded as appropriate, 19.4% (6/31) were graded as borderline, and 48.4% (15/31) were graded as inappropriate.</div></div><div><h3>Conclusions</h3><div>In this exploratory study, GPT-4 – a popular and publicly accessible chat-based AI model – was easily prompted to support CVD prevention misinformation. Misinformation-supporting arguments and scientific references were inappropriate due to inaccurate content and/or references nearly 50% of the time. Robust research efforts and policies are needed to study and prevent AI-enabled propagation of misinformation regarding CVD prevention.</div></div>","PeriodicalId":72173,"journal":{"name":"American journal of preventive cardiology","volume":null,"pages":null},"PeriodicalIF":4.3000,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"EVALUATING MISINFORMATION REGARDING CARDIOVASCULAR DISEASE PREVENTION OBTAINED ON A POPULAR, PUBLICLY ACCESSIBLE ARTIFICIAL INTELLIGENCE MODEL (GPT-4)\",\"authors\":\"\",\"doi\":\"10.1016/j.ajpc.2024.100806\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><h3>Therapeutic Area</h3><div>Other: Artificial intelligence; Misinformation</div></div><div><h3>Background</h3><div>Misinformation regarding CVD prevention is prevalent on the internet and on social media. Chat-based artificial intelligence (AI) models such as ChatGPT have gained over 100 million users, are publicly accessible, and may provide appropriate information for simple CVD prevention topics. Whether these public AI models may propagate misinformation regarding CVD prevention is uncertain.</div></div><div><h3>Methods</h3><div>This study was performed in March 2024 using the subscription-based version of GPT-4 (OpenAI, USA). Prompts regarding six CVD prevention topics (statin therapy and muscle-side effects, dementia, and liver disease; fish oil; supplements; and low-density lipoprotein-cholesterol and heart disease) were posed. Prompts were framed in two tones: a neutral tone and a misinformation-prompting tone. The misinformation-prompting tone requested specific arguments and scientific references to support misinformation. Each tone and topic was prompted in a different chatbot instance. Each response was reviewed by a board-certified cardiologist specializing in preventive cardiology at a tertiary care center. If a response had multiple bullet-points with individual scientific references, each bullet-point was graded separately. Responses were graded as appropriate (accurate content and references), borderline (minor inaccuracies or references published >20 years ago), or inappropriate (inaccurate content and/or references, including non-existent references).</div></div><div><h3>Results</h3><div>For the six prompts posed with a neutral tone, all responses lacked scientific references and were graded as appropriate (100%). For all six prompts posed with a misinformation-prompting tone, each response consisted of multiple discrete bullet-points with a scientific reference for each individual point. Of 31 bullet-points across the six topics obtained using a misinformation-prompting tone, 32.2% (10/31) were graded as appropriate, 19.4% (6/31) were graded as borderline, and 48.4% (15/31) were graded as inappropriate.</div></div><div><h3>Conclusions</h3><div>In this exploratory study, GPT-4 – a popular and publicly accessible chat-based AI model – was easily prompted to support CVD prevention misinformation. Misinformation-supporting arguments and scientific references were inappropriate due to inaccurate content and/or references nearly 50% of the time. Robust research efforts and policies are needed to study and prevent AI-enabled propagation of misinformation regarding CVD prevention.</div></div>\",\"PeriodicalId\":72173,\"journal\":{\"name\":\"American journal of preventive cardiology\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":4.3000,\"publicationDate\":\"2024-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"American journal of preventive cardiology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2666667724001740\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"CARDIAC & CARDIOVASCULAR SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"American journal of preventive cardiology","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666667724001740","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CARDIAC & CARDIOVASCULAR SYSTEMS","Score":null,"Total":0}
EVALUATING MISINFORMATION REGARDING CARDIOVASCULAR DISEASE PREVENTION OBTAINED ON A POPULAR, PUBLICLY ACCESSIBLE ARTIFICIAL INTELLIGENCE MODEL (GPT-4)
Therapeutic Area
Other: Artificial intelligence; Misinformation
Background
Misinformation regarding CVD prevention is prevalent on the internet and on social media. Chat-based artificial intelligence (AI) models such as ChatGPT have gained over 100 million users, are publicly accessible, and may provide appropriate information for simple CVD prevention topics. Whether these public AI models may propagate misinformation regarding CVD prevention is uncertain.
Methods
This study was performed in March 2024 using the subscription-based version of GPT-4 (OpenAI, USA). Prompts regarding six CVD prevention topics (statin therapy and muscle-side effects, dementia, and liver disease; fish oil; supplements; and low-density lipoprotein-cholesterol and heart disease) were posed. Prompts were framed in two tones: a neutral tone and a misinformation-prompting tone. The misinformation-prompting tone requested specific arguments and scientific references to support misinformation. Each tone and topic was prompted in a different chatbot instance. Each response was reviewed by a board-certified cardiologist specializing in preventive cardiology at a tertiary care center. If a response had multiple bullet-points with individual scientific references, each bullet-point was graded separately. Responses were graded as appropriate (accurate content and references), borderline (minor inaccuracies or references published >20 years ago), or inappropriate (inaccurate content and/or references, including non-existent references).
Results
For the six prompts posed with a neutral tone, all responses lacked scientific references and were graded as appropriate (100%). For all six prompts posed with a misinformation-prompting tone, each response consisted of multiple discrete bullet-points with a scientific reference for each individual point. Of 31 bullet-points across the six topics obtained using a misinformation-prompting tone, 32.2% (10/31) were graded as appropriate, 19.4% (6/31) were graded as borderline, and 48.4% (15/31) were graded as inappropriate.
Conclusions
In this exploratory study, GPT-4 – a popular and publicly accessible chat-based AI model – was easily prompted to support CVD prevention misinformation. Misinformation-supporting arguments and scientific references were inappropriate due to inaccurate content and/or references nearly 50% of the time. Robust research efforts and policies are needed to study and prevent AI-enabled propagation of misinformation regarding CVD prevention.