Battle of the authors: Comparing neurosurgery articles written by humans and AI

IF 1.9 4区 医学 Q3 CLINICAL NEUROLOGY Journal of Clinical Neuroscience Pub Date : 2025-02-25 DOI:10.1016/j.jocn.2025.111152
Mehmet Yigit Akgun , Melihcan Savasci , Caner Gunerbuyuk , Sezer Onur Gunara , Tunc Oktenoglu , Ali Fahir Ozer , Ozkan Ates
{"title":"Battle of the authors: Comparing neurosurgery articles written by humans and AI","authors":"Mehmet Yigit Akgun ,&nbsp;Melihcan Savasci ,&nbsp;Caner Gunerbuyuk ,&nbsp;Sezer Onur Gunara ,&nbsp;Tunc Oktenoglu ,&nbsp;Ali Fahir Ozer ,&nbsp;Ozkan Ates","doi":"10.1016/j.jocn.2025.111152","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><div>The advancement of artificial intelligence (AI) has led to its application in various fields, including medical literature. This study compares the quality of neurosurgery articles written by human authors and those generated by ChatGPT, an advanced AI model. The objective was to determine if AI-generated articles meet the standards of human-written academic papers.</div></div><div><h3>Methods</h3><div>A total of 10 neurosurgery articles, 5 written by humans and 5 by ChatGPT, were evaluated by a panel of blinded experts. The assessment parameters included overall impression, readability, criteria satisfaction, and degree of detail. Additionally, readability scores were calculated using the Lix score and the Flesch-Kincaid grade level. Preference and identification tests were also conducted to determine if experts could distinguish between the two types of articles.</div></div><div><h3>Results</h3><div>The study found no significant differences in the overall quality parameters between human-written and ChatGPT −generated articles. Readability scores were higher for ChatGPT articles (Lix score: 35 vs. 26, Flesch-Kincaid grade level: 10 vs. 8). Experts correctly identified the authorship of the articles 61% of the time, with preferences almost evenly split (47% preferred CHATGPT, 44% preferred human, and 9% had no preference). The most statistically significant result was the higher readability scores of CHATGPT-generated articles, indicating that AI can produce more readable content than human authors.</div></div><div><h3>Conclusion</h3><div>ChatGPT is capable of generating neurosurgery articles that are comparable in quality to those written by humans. The higher readability scores of AI-generated articles suggest that ChatGPT can enhance the accessibility of scientific literature. This study supports the potential integration of AI in academic writing, offering a valuable tool for researchers and medical professionals.</div></div>","PeriodicalId":15487,"journal":{"name":"Journal of Clinical Neuroscience","volume":"135 ","pages":"Article 111152"},"PeriodicalIF":1.9000,"publicationDate":"2025-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Clinical Neuroscience","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0967586825001249","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"CLINICAL NEUROLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Background

The advancement of artificial intelligence (AI) has led to its application in various fields, including medical literature. This study compares the quality of neurosurgery articles written by human authors and those generated by ChatGPT, an advanced AI model. The objective was to determine if AI-generated articles meet the standards of human-written academic papers.

Methods

A total of 10 neurosurgery articles, 5 written by humans and 5 by ChatGPT, were evaluated by a panel of blinded experts. The assessment parameters included overall impression, readability, criteria satisfaction, and degree of detail. Additionally, readability scores were calculated using the Lix score and the Flesch-Kincaid grade level. Preference and identification tests were also conducted to determine if experts could distinguish between the two types of articles.

Results

The study found no significant differences in the overall quality parameters between human-written and ChatGPT −generated articles. Readability scores were higher for ChatGPT articles (Lix score: 35 vs. 26, Flesch-Kincaid grade level: 10 vs. 8). Experts correctly identified the authorship of the articles 61% of the time, with preferences almost evenly split (47% preferred CHATGPT, 44% preferred human, and 9% had no preference). The most statistically significant result was the higher readability scores of CHATGPT-generated articles, indicating that AI can produce more readable content than human authors.

Conclusion

ChatGPT is capable of generating neurosurgery articles that are comparable in quality to those written by humans. The higher readability scores of AI-generated articles suggest that ChatGPT can enhance the accessibility of scientific literature. This study supports the potential integration of AI in academic writing, offering a valuable tool for researchers and medical professionals.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
求助全文
约1分钟内获得全文 去求助
来源期刊
Journal of Clinical Neuroscience
Journal of Clinical Neuroscience 医学-临床神经学
CiteScore
4.50
自引率
0.00%
发文量
402
审稿时长
40 days
期刊介绍: This International journal, Journal of Clinical Neuroscience, publishes articles on clinical neurosurgery and neurology and the related neurosciences such as neuro-pathology, neuro-radiology, neuro-ophthalmology and neuro-physiology. The journal has a broad International perspective, and emphasises the advances occurring in Asia, the Pacific Rim region, Europe and North America. The Journal acts as a focus for publication of major clinical and laboratory research, as well as publishing solicited manuscripts on specific subjects from experts, case reports and other information of interest to clinicians working in the clinical neurosciences.
期刊最新文献
Environmental risk factors of late-onset multiple sclerosis: A population-based case-control study Stereotactic radiosurgery for recurrent high-grade gliomas Compliance and factors affecting reporting of spinal stenosis and spinal-fusion-related clinical trials to ClinicalTrials.gov Battle of the authors: Comparing neurosurgery articles written by humans and AI Characterizing the complication profile of spinal robotic systems: A MAUDE analysis of device failures and associated complications by device manufacturer and brand name
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1