{"title":"Artificial Intelligence-Generated Writing in the ERAS Personal Statement: An Emerging Quandary for Post-graduate Medical Education.","authors":"Hugh Burke, Rebecca Kazinka, Raghu Gandhi, Aimee Murray","doi":"10.1007/s40596-024-02080-9","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>This study was designed to investigate if artificial intelligence (AI) detection software can determine the use of AI in personal statements for residency applications.</p><p><strong>Method: </strong>Previously written personal statements were collected from physicians who had already matched to residency through the Electronic Residency Application System. Physicians were recruited for the study through collegial relationships and were given study information via email. The study team constructed five parallel statements from the shared personal statements to prompt AI to create a personal statement of similar content. An online AI detection tool, GPTZero, was used to assess all the personal statements. Statistical analyses were conducted using R. Descriptive statistics, t-tests, and Pearson correlations were used to assess the data.</p><p><strong>Results: </strong>Eight physicians' statements were compared to eight AI-generated statements. GPTZero was able to correctly identify AI-generated writing, assigning them significantly higher AI probability scores compared to human-authored essays. Human-generated statements were considered more readable, used shorter words with fewer syllables, and had more sentences compared to AI-generated essays. Longer average sentence length, low readability scores, and high SAT word percentages were strongly associated with AI-generated essays.</p><p><strong>Conclusions: </strong>This study shows the capacity of GPTZero to distinguish human-created versus AI-generated writing. Use of AI can pose significant ethical challenges and carries a risk of inadvertent harm to certain applicants and erosion of trust in the application process. Authors suggest standardization of protocol regarding the use of AI prior to its integration in post-graduate medical education.</p>","PeriodicalId":7069,"journal":{"name":"Academic Psychiatry","volume":" ","pages":""},"PeriodicalIF":2.2000,"publicationDate":"2024-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Academic Psychiatry","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1007/s40596-024-02080-9","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION & EDUCATIONAL RESEARCH","Score":null,"Total":0}
引用次数: 0
Abstract
Objective: This study was designed to investigate if artificial intelligence (AI) detection software can determine the use of AI in personal statements for residency applications.
Method: Previously written personal statements were collected from physicians who had already matched to residency through the Electronic Residency Application System. Physicians were recruited for the study through collegial relationships and were given study information via email. The study team constructed five parallel statements from the shared personal statements to prompt AI to create a personal statement of similar content. An online AI detection tool, GPTZero, was used to assess all the personal statements. Statistical analyses were conducted using R. Descriptive statistics, t-tests, and Pearson correlations were used to assess the data.
Results: Eight physicians' statements were compared to eight AI-generated statements. GPTZero was able to correctly identify AI-generated writing, assigning them significantly higher AI probability scores compared to human-authored essays. Human-generated statements were considered more readable, used shorter words with fewer syllables, and had more sentences compared to AI-generated essays. Longer average sentence length, low readability scores, and high SAT word percentages were strongly associated with AI-generated essays.
Conclusions: This study shows the capacity of GPTZero to distinguish human-created versus AI-generated writing. Use of AI can pose significant ethical challenges and carries a risk of inadvertent harm to certain applicants and erosion of trust in the application process. Authors suggest standardization of protocol regarding the use of AI prior to its integration in post-graduate medical education.
期刊介绍:
Academic Psychiatry is the international journal of the American Association of Chairs of Departments of Psychiatry, American Association of Directors of Psychiatric Residency Training, Association for Academic Psychiatry, and Association of Directors of Medical Student Education in Psychiatry.
Academic Psychiatry publishes original, scholarly work in psychiatry and the behavioral sciences that focuses on innovative education, academic leadership, and advocacy.
The scope of the journal includes work that furthers knowledge and stimulates evidence-based advances in academic psychiatry in the following domains: education and training, leadership and administration, career and professional development, ethics and professionalism, and health and well-being.