Tianqi Tang , Jingrong Sha , Yanan Zhao , Saidi Wang , Zibin Wang , Sha Shen
{"title":"Unveiling the efficacy of ChatGPT in evaluating critical thinking skills through peer feedback analysis: Leveraging existing classification criteria","authors":"Tianqi Tang , Jingrong Sha , Yanan Zhao , Saidi Wang , Zibin Wang , Sha Shen","doi":"10.1016/j.tsc.2024.101607","DOIUrl":null,"url":null,"abstract":"<div><p>This study investigates the potential of using ChatGPT, a large language model, to assess students' critical thinking in online peer feedback. With the rapid development of technology, big language models, such as ChatGPT, have made significant progress in natural language processing in recent years and have good potential for application in teaching evaluation and feedback. However, can generative AI help educational practitioners in teaching and learning? How to accurately assess students' critical thinking using generative AI remains a challenging task. This study investigates whether ChatGPT can effectively evaluate critical thinking using established coding systems. By comparing the consistency and accuracy of manual coding with ChatGPT coding in online peer feedback texts, it clarifies how ChatGPT processes online peer feedback data and conducts assessments. Through a comprehensive analysis employing various metrics including precision, recall, F1 score, and a confusion matrix, we assess ChatGPT's performance. Additionally, we group students and analyze how ChatGPT's assessments relate to their critical thinking levels. Our findings suggest that the ChatGPT demonstrated some ability to assess higher dimensions of critical thinking, but showed limitations in assessing the more granular secondary dimensions under the higher dimensions of critical thinking. However for this kind of granular assessment will more accurately capture the level of learning critical thinking. Surprisingly, ChatGPT's evaluations are not influenced by students' critical thinking levels. This study underscores ChatGPT's potential in automating critical thinking assessment at scale, alleviating the burden on educators and enhancing understanding of critical thinking in peer feedback.</p></div>","PeriodicalId":47729,"journal":{"name":"Thinking Skills and Creativity","volume":"53 ","pages":"Article 101607"},"PeriodicalIF":3.7000,"publicationDate":"2024-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Thinking Skills and Creativity","FirstCategoryId":"95","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1871187124001457","RegionNum":2,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"Social Sciences","Score":null,"Total":0}
引用次数: 0
Abstract
This study investigates the potential of using ChatGPT, a large language model, to assess students' critical thinking in online peer feedback. With the rapid development of technology, big language models, such as ChatGPT, have made significant progress in natural language processing in recent years and have good potential for application in teaching evaluation and feedback. However, can generative AI help educational practitioners in teaching and learning? How to accurately assess students' critical thinking using generative AI remains a challenging task. This study investigates whether ChatGPT can effectively evaluate critical thinking using established coding systems. By comparing the consistency and accuracy of manual coding with ChatGPT coding in online peer feedback texts, it clarifies how ChatGPT processes online peer feedback data and conducts assessments. Through a comprehensive analysis employing various metrics including precision, recall, F1 score, and a confusion matrix, we assess ChatGPT's performance. Additionally, we group students and analyze how ChatGPT's assessments relate to their critical thinking levels. Our findings suggest that the ChatGPT demonstrated some ability to assess higher dimensions of critical thinking, but showed limitations in assessing the more granular secondary dimensions under the higher dimensions of critical thinking. However for this kind of granular assessment will more accurately capture the level of learning critical thinking. Surprisingly, ChatGPT's evaluations are not influenced by students' critical thinking levels. This study underscores ChatGPT's potential in automating critical thinking assessment at scale, alleviating the burden on educators and enhancing understanding of critical thinking in peer feedback.
期刊介绍:
Thinking Skills and Creativity is a new journal providing a peer-reviewed forum for communication and debate for the community of researchers interested in teaching for thinking and creativity. Papers may represent a variety of theoretical perspectives and methodological approaches and may relate to any age level in a diversity of settings: formal and informal, education and work-based.