Çağdaş Pamuk, Abdullah Faruk Uyanık, Ersin Kuyucu, Meriç Uğurlar
{"title":"Can ChatGPT pass the Turkish Orthopedics and Traumatology Board Examination? Turkish orthopedic surgeons versus artificial intelligence.","authors":"Çağdaş Pamuk, Abdullah Faruk Uyanık, Ersin Kuyucu, Meriç Uğurlar","doi":"10.14744/tjtes.2025.07724","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Artificial intelligence has been shown to achieve successful outcomes in various orthopedic qualification examinations worldwide. This study aims to assess the performance of ChatGPT in the written section of the Turkish Orthopedics and Traumatology Board Examination, compare its results with those of candidates who took the exam, and determine whether ChatGPT is sufficient to achieve a passing score.</p><p><strong>Methods: </strong>This retrospective observational study evaluated whether ChatGPT achieved a passing grade on 400 publicly available questions from the Turkish orthopedics qualification exam over the past four years. ChatGPT's performance was compared with the mean scores of the candidates who took the exam.</p><p><strong>Results: </strong>A total of 627 candidates participated in the four exams included in the study, of whom 292 (46.5%) passed. ChatGPT received higher scores than 619 (98.7%) of the candidates. In all exams conducted between 2020 and 2023, ChatGPT achieved significantly higher scores than the mean exam success rate (p=0.012, p=0.012, p=0.002, p=0.005, respectively). Of the 400 questions analyzed, 36 (9%) included figures.</p><p><strong>Conclusion: </strong>This is the first study to evaluate the performance of ChatGPT in the Turkish orthopedics proficiency exam. Our findings indicate that ChatGPT demonstrated high success in the Turkish Orthopedics and Traumatology Board Examination (TOTBE) written exam, achieving higher scores than the vast majority of candidates taking the exam (98.7%). ChatGPT performed well in the first part of the proficiency exam, where only theoretical knowledge is assessed. However, the human factor, which synthesizes both theoretical and practical knowledge, remains essential in daily medical practice.</p>","PeriodicalId":94263,"journal":{"name":"Ulusal travma ve acil cerrahi dergisi = Turkish journal of trauma & emergency surgery : TJTES","volume":"31 3","pages":"310-315"},"PeriodicalIF":0.0000,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ulusal travma ve acil cerrahi dergisi = Turkish journal of trauma & emergency surgery : TJTES","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.14744/tjtes.2025.07724","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Background: Artificial intelligence has been shown to achieve successful outcomes in various orthopedic qualification examinations worldwide. This study aims to assess the performance of ChatGPT in the written section of the Turkish Orthopedics and Traumatology Board Examination, compare its results with those of candidates who took the exam, and determine whether ChatGPT is sufficient to achieve a passing score.
Methods: This retrospective observational study evaluated whether ChatGPT achieved a passing grade on 400 publicly available questions from the Turkish orthopedics qualification exam over the past four years. ChatGPT's performance was compared with the mean scores of the candidates who took the exam.
Results: A total of 627 candidates participated in the four exams included in the study, of whom 292 (46.5%) passed. ChatGPT received higher scores than 619 (98.7%) of the candidates. In all exams conducted between 2020 and 2023, ChatGPT achieved significantly higher scores than the mean exam success rate (p=0.012, p=0.012, p=0.002, p=0.005, respectively). Of the 400 questions analyzed, 36 (9%) included figures.
Conclusion: This is the first study to evaluate the performance of ChatGPT in the Turkish orthopedics proficiency exam. Our findings indicate that ChatGPT demonstrated high success in the Turkish Orthopedics and Traumatology Board Examination (TOTBE) written exam, achieving higher scores than the vast majority of candidates taking the exam (98.7%). ChatGPT performed well in the first part of the proficiency exam, where only theoretical knowledge is assessed. However, the human factor, which synthesizes both theoretical and practical knowledge, remains essential in daily medical practice.