Generative AI: Potential and Pitfalls in Academic Publishing

Jason A. Silverman, Sabina A. Ali, Anna Rybak, Johannes B. van Goudoever, Neal S. Leleiko
{"title":"Generative AI: Potential and Pitfalls in Academic Publishing","authors":"Jason A. Silverman, Sabina A. Ali, Anna Rybak, Johannes B. van Goudoever, Neal S. Leleiko","doi":"10.1097/pg9.0000000000000387","DOIUrl":null,"url":null,"abstract":"Generative artificial intelligence (AI) has attracted enormous attention since the release of ChatGPT (1) in late 2022. Other generative AI chatbots used to generate text (including Bard, Claude, and CoPilot (2–4)) and images (including Dall-E, Midjourney, and Stable Diffusion (5–7)) have likewise seen a remarkable explosion in development and uptake. While ChatGPT has been promoted for its potential to assist users in efficiently and easily creating text to serve a wide range of purposes, this editorial focuses on the journal’s opinion on the use of generative AI technology in creating articles submitted to this journal. ChatGPT utilizes a technology known as a generative pretrained transformer large language model (8). This is a form of machine learning in which very large language-based data sets are used to train computers to comprehend natural language. While natural language processing is not new, ChatGPT is currently unique in its ability to not only understand queries and information to which it has access, but also to generate new, comprehensive, and fluent language-based content. Models, such as ChatGPT, which generates language, and Dall-E, Midjourney, or Stable Diffusion, which all generate images, are collectively referred to as “generative AI”, and can be seen as marking a dramatic shift in both the capabilities and widespread access to AI technology. Since its release on November 30, 2022, ChatGPT achieved the most rapid adoption of a consumer software application in history by amassing over 100 million users by January 2023 (9). Users have leveraged ChatGPT to write software, song lyrics, stories, poems, and letters. With its ability to recall previous prompts within a conversation, users can fine-tune its responses to modify the content or tone of the content it generates. The underlying model used to generate content is continuously improved through feedback from users. While users easily identified flaws in the responses provided by the earlier GPT-3 version, the model has iterated quickly. GPT-4, OpenAI’s latest effort in scaling up deep learning, has been described by its creators to exhibit “human-level performance on various professional and academic benchmarks” (10). It has demonstrated impressive abilities in passing standardized examinations including the Law School Admission Test, Scholastic Aptitude Test, a unified bar exam, and even the United States Medical Licensing Exam (10). ChatGPT, and other large language model-based applications, use large data sets consisting of text available on the web. Sources may include articles, books, web-based advertising, and social media posts. ChatGPT offers several exciting and desirable potential benefits, including its potential to make completing written articles more quickly as well as completing literature review summaries. With its ability to help users write in English fluently, it has been touted as a way to help make academic publishing more equitable and diverse (11). The time saved in summarizing data and generating articles could help researchers publish their studies faster, yielding more time to work on new experimental designs, grant applications, and more. This could significantly accelerate innovation and potentially lead to breakthroughs across many disciplines. We think this technology has enormous potential; however, important risks and limitations still exist that must also be acknowledged. Using ChatGPT or other language models may lead to inaccuracies, biases, and unintended plagiarism. It has been repeatedly shown that these models may create well-written text that has little relationship to reality in a phenomenon sometimes referred to as “hallucinations” (12,13). In particular, ChatGPT may generate text complete with appropriate-sounding but completely fictional citations included (14). It is obvious that articles containing factual inaccuracies, invented citations, and plagiarized content do not meet the editorial standards of our journals. Based on a review of the current capabilities and limitations of available generative AI software, we require our contributors to follow responsible and transparent practices and policies. It is necessary that authors ensure that references cited state what is attributed to them and that the overall written document is logical and consistent with the actual findings reported in the submission. In line with the International Committee of Medical Journal Editors, we expect each submission to reflect the expertise of the author or authors, who are all ultimately responsible for every word in their submission (15). As such, generative AI software, including ChatGPT and others, may not be listed as a coauthor on any submitted article. Authors must clearly state in their article the extent to which AI technologies were used in data analysis, literature review, and article preparation. This will aid reviewers and editors in checking potential biases, inaccuracies, and improper source attribution while providing readers with a transparent view of how the article was created. We ask that authors (and other interested parties) state in their letter of submission the extent to which AI was used as well as other related information they deem pertinent. This information will help ensure the integrity of the editorial review and publication process and will serve as a learning tool for the editors. We all have a great deal to learn about generative AI and its place in academic publishing. We need to evolve together along with the technical capabilities of AI. We expect to modify our editorial policies and instructions for authors often in the coming months and years as these tools, and our understanding of their impacts and best uses continue to develop.","PeriodicalId":17618,"journal":{"name":"JPGN Reports","volume":"236 2","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"JPGN Reports","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1097/pg9.0000000000000387","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Generative artificial intelligence (AI) has attracted enormous attention since the release of ChatGPT (1) in late 2022. Other generative AI chatbots used to generate text (including Bard, Claude, and CoPilot (2–4)) and images (including Dall-E, Midjourney, and Stable Diffusion (5–7)) have likewise seen a remarkable explosion in development and uptake. While ChatGPT has been promoted for its potential to assist users in efficiently and easily creating text to serve a wide range of purposes, this editorial focuses on the journal’s opinion on the use of generative AI technology in creating articles submitted to this journal. ChatGPT utilizes a technology known as a generative pretrained transformer large language model (8). This is a form of machine learning in which very large language-based data sets are used to train computers to comprehend natural language. While natural language processing is not new, ChatGPT is currently unique in its ability to not only understand queries and information to which it has access, but also to generate new, comprehensive, and fluent language-based content. Models, such as ChatGPT, which generates language, and Dall-E, Midjourney, or Stable Diffusion, which all generate images, are collectively referred to as “generative AI”, and can be seen as marking a dramatic shift in both the capabilities and widespread access to AI technology. Since its release on November 30, 2022, ChatGPT achieved the most rapid adoption of a consumer software application in history by amassing over 100 million users by January 2023 (9). Users have leveraged ChatGPT to write software, song lyrics, stories, poems, and letters. With its ability to recall previous prompts within a conversation, users can fine-tune its responses to modify the content or tone of the content it generates. The underlying model used to generate content is continuously improved through feedback from users. While users easily identified flaws in the responses provided by the earlier GPT-3 version, the model has iterated quickly. GPT-4, OpenAI’s latest effort in scaling up deep learning, has been described by its creators to exhibit “human-level performance on various professional and academic benchmarks” (10). It has demonstrated impressive abilities in passing standardized examinations including the Law School Admission Test, Scholastic Aptitude Test, a unified bar exam, and even the United States Medical Licensing Exam (10). ChatGPT, and other large language model-based applications, use large data sets consisting of text available on the web. Sources may include articles, books, web-based advertising, and social media posts. ChatGPT offers several exciting and desirable potential benefits, including its potential to make completing written articles more quickly as well as completing literature review summaries. With its ability to help users write in English fluently, it has been touted as a way to help make academic publishing more equitable and diverse (11). The time saved in summarizing data and generating articles could help researchers publish their studies faster, yielding more time to work on new experimental designs, grant applications, and more. This could significantly accelerate innovation and potentially lead to breakthroughs across many disciplines. We think this technology has enormous potential; however, important risks and limitations still exist that must also be acknowledged. Using ChatGPT or other language models may lead to inaccuracies, biases, and unintended plagiarism. It has been repeatedly shown that these models may create well-written text that has little relationship to reality in a phenomenon sometimes referred to as “hallucinations” (12,13). In particular, ChatGPT may generate text complete with appropriate-sounding but completely fictional citations included (14). It is obvious that articles containing factual inaccuracies, invented citations, and plagiarized content do not meet the editorial standards of our journals. Based on a review of the current capabilities and limitations of available generative AI software, we require our contributors to follow responsible and transparent practices and policies. It is necessary that authors ensure that references cited state what is attributed to them and that the overall written document is logical and consistent with the actual findings reported in the submission. In line with the International Committee of Medical Journal Editors, we expect each submission to reflect the expertise of the author or authors, who are all ultimately responsible for every word in their submission (15). As such, generative AI software, including ChatGPT and others, may not be listed as a coauthor on any submitted article. Authors must clearly state in their article the extent to which AI technologies were used in data analysis, literature review, and article preparation. This will aid reviewers and editors in checking potential biases, inaccuracies, and improper source attribution while providing readers with a transparent view of how the article was created. We ask that authors (and other interested parties) state in their letter of submission the extent to which AI was used as well as other related information they deem pertinent. This information will help ensure the integrity of the editorial review and publication process and will serve as a learning tool for the editors. We all have a great deal to learn about generative AI and its place in academic publishing. We need to evolve together along with the technical capabilities of AI. We expect to modify our editorial policies and instructions for authors often in the coming months and years as these tools, and our understanding of their impacts and best uses continue to develop.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
生成人工智能:学术出版的潜力和陷阱
这将有助于审稿人和编辑检查潜在的偏见、不准确和不适当的来源归因,同时为读者提供一个透明的视角,了解文章是如何创建的。我们要求作者(和其他相关方)在投稿信中说明人工智能的使用程度以及他们认为相关的其他相关信息。这些信息将有助于确保编辑审查和出版过程的完整性,并将作为编辑的学习工具。关于生成式人工智能及其在学术出版中的地位,我们都有很多需要学习的地方。我们需要与人工智能的技术能力一起发展。随着这些工具以及我们对它们的影响和最佳用途的理解不断发展,我们希望在未来的几个月和几年中经常修改我们的编辑政策和作者指导。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
‘You can be as vigilant as you can, yet they make their way in’: A descriptive study of parent and caregiver perspectives towards keeping children safe from button batteries Co‐occurrence of collagenous gastrointestinal disease in siblings in early childhood: New insight into a rare condition Impact of gender, race, and age of onset on the phenotype and comorbidities of pediatric eosinophilic esophagitis Pediatric eosinophilic gastritis treated with benralizumab: A case report B cell depletion for autoimmune liver diseases: A retrospective review of indications and outcomes
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1