R. Garg, V. L. Urs, Akshya Anand Agrawal, Sarvesh Kumar Chaudhary, V. Paliwal, Sujita Kumar Kar
{"title":"Exploring the role of ChatGPT in patient care (diagnosis and treatment) and medical research: A systematic review","authors":"R. Garg, V. L. Urs, Akshya Anand Agrawal, Sarvesh Kumar Chaudhary, V. Paliwal, Sujita Kumar Kar","doi":"10.1101/2023.06.13.23291311","DOIUrl":null,"url":null,"abstract":"Background ChatGPT(Chat Generative Pre-trained Transformer) is an artificial intelligence (AI) based on a natural language processing tool developed by OpenAI (California, USA). This systematic review examines the potential of Chat GPT in diagnosing and treating patients and its contributions to medical research. Methods In order to locate articles on ChatGPT's use in clinical practise and medical research, this systematic review used PRISMA standards and conducted database searches across several sources. Selected records were analysed using ChatGPT, which also produced a summary for each article. The resultant word document was transformed to a PDF and handled using ChatPDF. The review looked at topics pertaining to scholarly publishing, clinical practise, and medical research. Results We reviewed 118 publications. There are difficulties and moral conundrums associated with using ChatGPT in therapeutic settings and medical research. Patient inquiries, note writing, decision-making, trial enrolment, data management, decision support, research support, and patient education are all things that ChatGPT can help with. However, the solutions it provides are frequently inadequate and inconsistent, presenting issues with its originality, privacy, accuracy, bias, and legality. When utilising ChatGPT for academic writings, there are issues with prejudice and plagiarism, and because it lacks human-like characteristics, its authority as an author is called into question. Conclusions ChatGPT has limitations when used in research and healthcare. Even while it aids in patient treatment, concerns regarding accuracy, authorship, and bias arise. Currently, ChatGPT can serve as a \"clinical assistant\" and be a huge assistance with research and scholarly writing.","PeriodicalId":46588,"journal":{"name":"Health Promotion Perspectives","volume":"13 1","pages":"183 - 191"},"PeriodicalIF":2.4000,"publicationDate":"2023-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Health Promotion Perspectives","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1101/2023.06.13.23291311","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"PUBLIC, ENVIRONMENTAL & OCCUPATIONAL HEALTH","Score":null,"Total":0}
引用次数: 6
Abstract
Background ChatGPT(Chat Generative Pre-trained Transformer) is an artificial intelligence (AI) based on a natural language processing tool developed by OpenAI (California, USA). This systematic review examines the potential of Chat GPT in diagnosing and treating patients and its contributions to medical research. Methods In order to locate articles on ChatGPT's use in clinical practise and medical research, this systematic review used PRISMA standards and conducted database searches across several sources. Selected records were analysed using ChatGPT, which also produced a summary for each article. The resultant word document was transformed to a PDF and handled using ChatPDF. The review looked at topics pertaining to scholarly publishing, clinical practise, and medical research. Results We reviewed 118 publications. There are difficulties and moral conundrums associated with using ChatGPT in therapeutic settings and medical research. Patient inquiries, note writing, decision-making, trial enrolment, data management, decision support, research support, and patient education are all things that ChatGPT can help with. However, the solutions it provides are frequently inadequate and inconsistent, presenting issues with its originality, privacy, accuracy, bias, and legality. When utilising ChatGPT for academic writings, there are issues with prejudice and plagiarism, and because it lacks human-like characteristics, its authority as an author is called into question. Conclusions ChatGPT has limitations when used in research and healthcare. Even while it aids in patient treatment, concerns regarding accuracy, authorship, and bias arise. Currently, ChatGPT can serve as a "clinical assistant" and be a huge assistance with research and scholarly writing.