首页 > 最新文献

Computers and Education Artificial Intelligence最新文献

英文 中文
Epistemic network analysis of in-service teachers’ competency to teach artificial intelligence for secondary education 在职教师中等教育人工智能教学能力的认知网络分析
Q1 Social Sciences Pub Date : 2025-12-01 DOI: 10.1016/j.caeai.2025.100520
King Woon Yau , Tianle Dong , Ching Sing Chai , Thomas K.F. Chiu , Helen Meng , Irwin King , Savio W.H. Wong , Yeung Yam
Teachers play a vital role in driving successful artificial intelligence (AI) education. Research on teachers' competency to teach AI (TCAI) is still limited. This study investigated the progression of in-service teachers' AI competency with the Technological Pedagogical Content Knowledge (TPACK) framework using Epistemic Network Analysis (ENA). Seven secondary school teachers who engaged in an AI education project were interviewed over a three-year period of curriculum development and implementation. The differences in ENA patterns in various stages indicated an evolution of teachers’ TPACK over the years. The ENA results also revealed different patterns between experienced and less experienced teachers. Experienced teachers tend to integrate their TPACK components with pedagogical considerations, whereas less experienced teachers focus more on content-related elements. The differences in ENA patterns indicate distinct progression paths with different focuses, highlighting the need to tailor professional development activities for different groups of teachers at various stages. These findings underscore the importance of continuous support and targeted training to enhance teachers' AI competency in AI education.
教师在推动人工智能(AI)教育成功方面发挥着至关重要的作用。关于教师人工智能(TCAI)教学能力的研究仍然有限。本研究运用认知网络分析(ENA),在技术教学内容知识(TPACK)框架下,对在职教师人工智能能力的发展进行了研究。参与人工智能教育项目的七位中学教师接受了为期三年的课程开发和实施采访。不同阶段ENA模式的差异反映了教师TPACK的演变过程。ENA的结果还揭示了经验丰富和经验不足的教师之间的不同模式。经验丰富的教师倾向于将他们的TPACK组件与教学考虑相结合,而经验不足的教师则更多地关注与内容相关的元素。ENA模式的差异表明了不同的发展路径和不同的重点,突出了在不同阶段为不同教师群体量身定制专业发展活动的必要性。这些发现强调了在人工智能教育中,持续支持和有针对性的培训对于提高教师的人工智能能力的重要性。
{"title":"Epistemic network analysis of in-service teachers’ competency to teach artificial intelligence for secondary education","authors":"King Woon Yau ,&nbsp;Tianle Dong ,&nbsp;Ching Sing Chai ,&nbsp;Thomas K.F. Chiu ,&nbsp;Helen Meng ,&nbsp;Irwin King ,&nbsp;Savio W.H. Wong ,&nbsp;Yeung Yam","doi":"10.1016/j.caeai.2025.100520","DOIUrl":"10.1016/j.caeai.2025.100520","url":null,"abstract":"<div><div>Teachers play a vital role in driving successful artificial intelligence (AI) education. Research on teachers' competency to teach AI (TCAI) is still limited. This study investigated the progression of in-service teachers' AI competency with the Technological Pedagogical Content Knowledge (TPACK) framework using Epistemic Network Analysis (ENA). Seven secondary school teachers who engaged in an AI education project were interviewed over a three-year period of curriculum development and implementation. The differences in ENA patterns in various stages indicated an evolution of teachers’ TPACK over the years. The ENA results also revealed different patterns between experienced and less experienced teachers. Experienced teachers tend to integrate their TPACK components with pedagogical considerations, whereas less experienced teachers focus more on content-related elements. The differences in ENA patterns indicate distinct progression paths with different focuses, highlighting the need to tailor professional development activities for different groups of teachers at various stages. These findings underscore the importance of continuous support and targeted training to enhance teachers' AI competency in AI education.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"9 ","pages":"Article 100520"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145681342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Developing a theory-grounded AI tool for the generation of culturally responsive lesson plans 开发一种基于理论的人工智能工具,用于生成与文化相关的课程计划
Q1 Social Sciences Pub Date : 2025-12-01 DOI: 10.1016/j.caeai.2025.100474
Matthew Nyaaba , Xiaoming Zhai
As educators begin using Generative AI (GenAI) for lesson planning, they often encounter generated content that fails to consider the classroom's cultural context. In this study, we address this issue by adopting a design science research approach to develop a theory-based prompt grounded in culturally responsive pedagogy (CRP) and using it to customize a Culturally Responsive Lesson Planner (CRLP) GPT. Guided by the CRP framework, the CRLP uses an Interactive Semi-Automated (ISA) prompt architecture that engages teachers in dialogue to collect cultural and contextual details before generating a lesson plan. To evaluate the CRLP's effectiveness, we asked two expert reviewers to compare Grade 7 “States of Matter” lesson plans for Ghana's Ashanti Region, generated with both the CRLP and the base GPT-4o using a standard prompt. The expert reviewers rated the CRLP-generated lesson plan higher in cultural elements identified (36 vs. 21 elements), accuracy (1.8 vs. 1.2), and curriculum relevance (2.0 vs. 1.3) than that created by the standard prompt within the base GPT-4o. The CRLP-generated lesson plan also included more Asante Twi examples such as “Solid” (ɛpono [wooden furniture], dadeɛ [metal objects], aboɔ [stones], and ntadeɛ [clothing]), recommended local teaching resources, and allowed teachers to make final revisions before generating the complete lesson plan. Additionally, the CRLP included the developer's contact details to encourage ongoing feedback and improvement. However, cultural hallucinations were slightly higher (0.75 vs. 0.5) in the CRLP-generated lesson plan compared with the standard GPT-4o prompt. These findings suggest that a GenAI tool grounded in educational theory is more effective in supporting the goals of education than the standard version. Furthermore, the CRLP and its ISA prompt strategy represent Human-in-the-loop system that has the potential to enhance teachers' AI literacy and culturally responsive pedagogy as they engage with the tool. We recommend future studies comparing CRLP and human-generated lesson plans, as well as empirical research that tests CRLP lesson plans in classroom settings.
随着教育工作者开始使用生成式人工智能(GenAI)进行课程规划,他们经常会遇到无法考虑课堂文化背景的生成内容。在本研究中,我们通过采用设计科学研究方法来开发基于文化响应教学法(CRP)的理论提示,并使用它来定制文化响应课程计划(CRLP) GPT,从而解决了这个问题。在CRP框架的指导下,CRLP使用交互式半自动(ISA)提示架构,让教师参与对话,在生成课程计划之前收集文化和上下文细节。为了评估CRLP的有效性,我们请了两位专家审稿人比较了加纳阿散蒂地区七年级的“物质状态”课程计划,这些课程计划是用CRLP和基础gpt - 40生成的,使用标准提示。专家评审员认为,与基础gpt - 40中的标准提示创建的教案相比,crlp生成的教案在识别的文化要素(36比21)、准确性(1.8比1.2)和课程相关性(2.0比1.3)方面都更高。该项目生成的课程计划还包括更多的Asante Twi例子,如“Solid”([木制家具]、[金属物品]、[石头]和[衣服]),推荐了当地的教学资源,并允许教师在生成完整的课程计划之前进行最后的修改。此外,CRLP还包括开发人员的联系方式,以鼓励持续的反馈和改进。然而,与标准的gpt - 40提示相比,crlp生成的课程计划中的文化幻觉略高(0.75 vs. 0.5)。这些发现表明,基于教育理论的GenAI工具在支持教育目标方面比标准版本更有效。此外,CRLP及其ISA提示策略代表了人在循环系统,有可能在教师使用该工具时提高他们的人工智能素养和文化响应教学法。我们建议未来的研究比较CRLP和人工生成的课程计划,以及在课堂环境中测试CRLP课程计划的实证研究。
{"title":"Developing a theory-grounded AI tool for the generation of culturally responsive lesson plans","authors":"Matthew Nyaaba ,&nbsp;Xiaoming Zhai","doi":"10.1016/j.caeai.2025.100474","DOIUrl":"10.1016/j.caeai.2025.100474","url":null,"abstract":"<div><div>As educators begin using Generative AI (GenAI) for lesson planning, they often encounter generated content that fails to consider the classroom's cultural context. In this study, we address this issue by adopting a design science research approach to develop a theory-based prompt grounded in culturally responsive pedagogy (CRP) and using it to customize a <em>Culturally Responsive Lesson Planner (CRLP) GPT.</em> Guided by the CRP framework, the CRLP uses an <em>Interactive</em> <em>S</em><em>emi-</em><em>Automated</em> (ISA) prompt architecture that engages teachers in dialogue to collect cultural and contextual details before generating a lesson plan. To evaluate the CRLP's effectiveness, we asked two expert reviewers to compare Grade 7 “States of Matter” lesson plans for Ghana's Ashanti Region, generated with both the CRLP and the base GPT-4o using a standard prompt. The expert reviewers rated the CRLP-generated lesson plan higher in cultural elements identified (36 vs. 21 elements), accuracy (1.8 vs. 1.2), and curriculum relevance (2.0 vs. 1.3) than that created by the standard prompt within the base GPT-4o. The CRLP-generated lesson plan also included more Asante Twi examples such as “Solid” (<em>ɛpono</em> [wooden furniture], <em>dadeɛ</em> [metal objects], <em>aboɔ</em> [stones], and <em>ntadeɛ</em> [clothing]), recommended local teaching resources, and allowed teachers to make final revisions before generating the complete lesson plan. Additionally, the CRLP included the developer's contact details to encourage ongoing feedback and improvement. However, cultural hallucinations were slightly higher (0.75 vs. 0.5) in the CRLP-generated lesson plan compared with the standard GPT-4o prompt. These findings suggest that a GenAI tool grounded in educational theory is more effective in supporting the goals of education than the standard version. Furthermore, the CRLP and its ISA prompt strategy represent Human-in-the-loop system that has the potential to enhance teachers' AI literacy and culturally responsive pedagogy as they engage with the tool. We recommend future studies comparing CRLP and human-generated lesson plans, as well as empirical research that tests CRLP lesson plans in classroom settings.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"9 ","pages":"Article 100474"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145681348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards responsible AI for education: Hybrid human-AI to confront the elephant in the room 走向负责任的人工智能教育:人类与人工智能的混合,以面对房间里的大象
Q1 Social Sciences Pub Date : 2025-12-01 DOI: 10.1016/j.caeai.2025.100524
Danial Hooshyar , Gustav Šír , Yeongwook Yang , Eve Kikas , Raija Hämäläinen , Tommi Kärkkäinen , Dragan Gašević , Roger Azevedo
Despite significant advancements in AI-driven educational systems and ongoing calls for responsible AI for education, several critical issues remain unresolved—acting as elephant in the room within AI in education, learning analytics, educational data mining, learning sciences, and educational psychology communities. This critical analysis identifies and examines nine persistent challenges across the conceptual, methodological, and ethical dimensions that continue to undermine the fairness, transparency, and effectiveness of current AI methods and applications in education. These include: 1) the lack of clarity around what AI for education truly means—often ignoring the distinct purposes, strengths, and limitations of different AI families—and the trend of equating it with domain-agnostic, company-driven large language models; 2) the widespread neglect of essential learning processes such as motivation, emotion, and (meta)cognition in AI-driven learner modelling and their contextual nature; 3) limited integration of domain knowledge and lack of stakeholder involvement in AI design and development; 4) continued use of non-sequential machine learning models on temporal educational data; 5) misuse of non-sequential metrics to evaluate sequential models; 6) using unreliable explainable AI methods to provide explanations for black-box models; 7) ignoring ethical guidelines in addressing data inconsistencies during model training; 8) use of mainstream AI methods for pattern discovery and learning analytics without systematic benchmarking; and 9) overemphasis on global prescriptions while overlooking localized, student-specific recommendations. Supported by theoretical and empirical research, we demonstrate how hybrid AI methods—specifically neural-symbolic AI—can address the elephant in the room and serve as the foundation for responsible, trustworthy AI systems in education.
尽管人工智能驱动的教育系统取得了重大进展,并且不断呼吁对教育负责任的人工智能,但仍有几个关键问题尚未解决——在教育、学习分析、教育数据挖掘、学习科学和教育心理学社区的人工智能领域,这些问题就像房间里的大象一样。这一批判性分析确定并考察了概念、方法和道德层面上的九个持续挑战,这些挑战继续破坏当前人工智能方法和教育应用的公平性、透明度和有效性。这些问题包括:1)教育领域人工智能的真正含义缺乏明确性——经常忽视不同人工智能家族的独特目的、优势和局限性——以及将其等同于领域不可知、公司驱动的大型语言模型的趋势;2)在人工智能驱动的学习者建模及其上下文性质中,普遍忽视了基本的学习过程,如动机、情感和(元)认知;3)领域知识整合有限,缺乏利益相关者参与人工智能设计和开发;4)在时序教育数据上继续使用非顺序机器学习模型;5)误用非顺序度量来评价顺序模型;6)使用不可靠的可解释AI方法为黑箱模型提供解释;7)在模型训练过程中忽视处理数据不一致的道德准则;8)使用主流人工智能方法进行模式发现和学习分析,而没有进行系统的基准测试;9)过分强调全球处方,而忽视了本地化的、针对学生的建议。在理论和实证研究的支持下,我们展示了混合人工智能方法——特别是神经符号人工智能——如何解决房间里的大象,并作为负责任的、值得信赖的教育人工智能系统的基础。
{"title":"Towards responsible AI for education: Hybrid human-AI to confront the elephant in the room","authors":"Danial Hooshyar ,&nbsp;Gustav Šír ,&nbsp;Yeongwook Yang ,&nbsp;Eve Kikas ,&nbsp;Raija Hämäläinen ,&nbsp;Tommi Kärkkäinen ,&nbsp;Dragan Gašević ,&nbsp;Roger Azevedo","doi":"10.1016/j.caeai.2025.100524","DOIUrl":"10.1016/j.caeai.2025.100524","url":null,"abstract":"<div><div>Despite significant advancements in AI-driven educational systems and ongoing calls for responsible AI for education, several critical issues remain unresolved—acting as <em>elephant in the room</em> within AI in education, learning analytics, educational data mining, learning sciences, and educational psychology communities. This critical analysis identifies and examines nine persistent challenges across the conceptual, methodological, and ethical dimensions that continue to undermine the fairness, transparency, and effectiveness of current AI methods and applications in education. These include: 1) the lack of clarity around what AI for education truly means—often ignoring the distinct purposes, strengths, and limitations of different AI families—and the trend of equating it with domain-agnostic, company-driven large language models; 2) the widespread neglect of essential learning processes such as motivation, emotion, and (meta)cognition in AI-driven learner modelling and their contextual nature; 3) limited integration of domain knowledge and lack of stakeholder involvement in AI design and development; 4) continued use of non-sequential machine learning models on temporal educational data; 5) misuse of non-sequential metrics to evaluate sequential models; 6) using unreliable explainable AI methods to provide explanations for black-box models; 7) ignoring ethical guidelines in addressing data inconsistencies during model training; 8) use of mainstream AI methods for pattern discovery and learning analytics without systematic benchmarking; and 9) overemphasis on global prescriptions while overlooking localized, student-specific recommendations. Supported by theoretical and empirical research, we demonstrate how hybrid AI methods—specifically neural-symbolic AI—can address the <em>elephant in the room</em> and serve as the foundation for responsible, trustworthy AI systems in education.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"9 ","pages":"Article 100524"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145681351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Is ChatGPT a good study companion? The role of AI-generated summaries and reflective prompts in learning from educational videos ChatGPT是一个好的学习伙伴吗?人工智能生成的摘要和反思性提示在学习教育视频中的作用
Q1 Social Sciences Pub Date : 2025-12-01 DOI: 10.1016/j.caeai.2025.100512
Ayşe Candan Şimşek , Gerrit Anders , Jonathan Göth , Luisa Specht , Markus Huff
Online videos have become a central tool in modern education. Alongside this shift, Artificial Intelligence (AI) is reshaping personalized learning experiences, with generative large language models like ChatGPT offering new ways to tailor information to individual learners. Based on the Cognitive Theory of Multimedia Learning (CTML), which proposes two principles that relate to the interaction with the learning material (segmenting and generative activity), we conducted two experiments in which participants were asked to pause an educational video at times of comprehension difficulty. In Experiment 1 (N = 101), we examined whether GPT-generated summaries -introduced at self-paced pause points-result in better learning compared to video transcripts. In Experiment 2 (N = 215), we compared the role of GPT-generated summaries and GPT-generated reflective prompts. Those elicited open-ended answers from the participants. We measured retention and transfer learning, as well as mental effort, and perceived task difficulty. Contrary to our expectations, we observed no differences between AI summaries and transcripts in terms of retention and transfer outcomes. Participants showed a learning effect indicating more correct answers after watching the video, but this effect did not differ between conditions. We can especially note that the motivation to engage in the material, as well as the difficulty and length of the video, may have affected the results. As research investigating the role of AI in educational settings is still new, future research can delve into finding the optimal conditions under which AI can benefit learning outcomes.
在线视频已经成为现代教育的核心工具。除了这种转变,人工智能(AI)正在重塑个性化学习体验,ChatGPT等生成式大型语言模型为个性化学习者提供了定制信息的新方法。基于多媒体学习的认知理论(CTML),提出了与学习材料互动相关的两个原则(分段和生成活动),我们进行了两个实验,要求参与者在理解困难时暂停教育视频。在实验1 (N = 101)中,我们检验了gpt生成的摘要(在自定节奏的暂停点引入)是否比视频成绩单产生更好的学习效果。在实验2 (N = 215)中,我们比较了gpt生成的摘要和gpt生成的反思提示的作用。这些问题引出了参与者的开放式回答。我们测量了保留和迁移学习,以及心理努力和感知任务难度。与我们的预期相反,我们观察到AI摘要和转录本在保留和转移结果方面没有差异。参与者在观看视频后表现出学习效果,表明更多的正确答案,但这种效果在不同条件下没有差异。我们可以特别注意到,参与材料的动机,以及视频的难度和长度,可能会影响结果。由于调查人工智能在教育环境中的作用的研究仍然是新的,未来的研究可以深入研究寻找人工智能可以促进学习成果的最佳条件。
{"title":"Is ChatGPT a good study companion? The role of AI-generated summaries and reflective prompts in learning from educational videos","authors":"Ayşe Candan Şimşek ,&nbsp;Gerrit Anders ,&nbsp;Jonathan Göth ,&nbsp;Luisa Specht ,&nbsp;Markus Huff","doi":"10.1016/j.caeai.2025.100512","DOIUrl":"10.1016/j.caeai.2025.100512","url":null,"abstract":"<div><div>Online videos have become a central tool in modern education. Alongside this shift, Artificial Intelligence (AI) is reshaping personalized learning experiences, with generative large language models like ChatGPT offering new ways to tailor information to individual learners. Based on the Cognitive Theory of Multimedia Learning (CTML), which proposes two principles that relate to the interaction with the learning material (segmenting and generative activity), we conducted two experiments in which participants were asked to pause an educational video at times of comprehension difficulty. In Experiment 1 (<em>N</em> = 101), we examined whether GPT-generated summaries -introduced at self-paced pause points-result in better learning compared to video transcripts. In Experiment 2 (<em>N</em> = 215), we compared the role of GPT-generated summaries and GPT-generated reflective prompts. Those elicited open-ended answers from the participants. We measured retention and transfer learning, as well as mental effort, and perceived task difficulty. Contrary to our expectations, we observed no differences between AI summaries and transcripts in terms of retention and transfer outcomes. Participants showed a learning effect indicating more correct answers after watching the video, but this effect did not differ between conditions. We can especially note that the motivation to engage in the material, as well as the difficulty and length of the video, may have affected the results. As research investigating the role of AI in educational settings is still new, future research can delve into finding the optimal conditions under which AI can benefit learning outcomes.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"9 ","pages":"Article 100512"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145614497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Role of online assessment system in formative evaluation of programming education 在线评价系统在程序设计教育形成性评价中的作用
Q1 Social Sciences Pub Date : 2025-12-01 DOI: 10.1016/j.caeai.2025.100515
Haitang Wan
With the rapid development of educational informatization, the IntelliAssessment is increasingly widely used in the formative evaluation of course teaching. Scalability testing (Python Locust framework) showed 100 % response rate under 500 concurrent requests (consistent with typical university course sizes), while 92 % response rate was observed at 1000 concurrent requests (an extreme stress test scenario). Security validation included a 94 % attack-blocking rate in penetration testing and 91.0 % F1-score for AI-driven phishing detection. The <2-s real-time feedback window (p50 = 1.2 s, p90 = 1.8 s, p99 = 2.3 s) is maintained for 90 % of interactions under typical loads, with latency degrading only at very high concurrency—pedagogically, this ensures timely formative feedback for most classroom scenarios. A supplementary analysis discussing current security limitations and the evolving nature of security threats has been added, along with potential development ideas for enhancing system security. These improvements aim to strengthen the comprehensiveness and scientific reliability of our manuscript. Statistics show that 27.65 % of the students who participated in the evaluation were very satisfied with the feedback, while 17.4 % thought that the feedback was helpful. As for understanding the assessment content, 3.15 % of the students indicated that they needed more clarity, indicating that the clarity of the assessment questions still required improvement. Among the students’ learning performance, 67.24 % scored higher than the passing line, indicating that most students can master the course content.
随着教育信息化的快速发展,智能评估在课程教学形成性评价中的应用越来越广泛。可伸缩性测试(Python Locust框架)显示,在500个并发请求(与典型的大学课程规模一致)下,响应率为100%,而在1000个并发请求(极端压力测试场景)下,响应率为92%。安全验证包括渗透测试中94%的攻击阻止率和人工智能驱动的网络钓鱼检测中91.0%的f1得分。2秒的实时反馈窗口(p50 = 1.2秒,p90 = 1.8秒,p99 = 2.3秒)在典型负载下保持90%的交互,延迟仅在非常高的并发性下降低-从教学角度来说,这确保了大多数课堂场景的及时形成反馈。本文还添加了一个补充分析,讨论了当前的安全限制和安全威胁的演变性质,以及增强系统安全性的潜在开发思想。这些改进旨在加强我们稿件的全面性和科学可靠性。统计显示,27.65%的参与评价的学生对反馈非常满意,17.4%的学生认为反馈有帮助。在对评估内容的理解上,有3.15%的学生表示需要更加清晰,说明评估问题的清晰性仍有待提高。在学生的学习成绩中,67.24%的学生得分高于及格线,说明大部分学生能够掌握课程内容。
{"title":"Role of online assessment system in formative evaluation of programming education","authors":"Haitang Wan","doi":"10.1016/j.caeai.2025.100515","DOIUrl":"10.1016/j.caeai.2025.100515","url":null,"abstract":"<div><div>With the rapid development of educational informatization, the IntelliAssessment is increasingly widely used in the formative evaluation of course teaching. Scalability testing (Python Locust framework) showed 100 % response rate under 500 concurrent requests (consistent with typical university course sizes), while 92 % response rate was observed at 1000 concurrent requests (an extreme stress test scenario). Security validation included a 94 % attack-blocking rate in penetration testing and 91.0 % F1-score for AI-driven phishing detection. The &lt;2-s real-time feedback window (p50 = 1.2 s, p90 = 1.8 s, p99 = 2.3 s) is maintained for 90 % of interactions under typical loads, with latency degrading only at very high concurrency—pedagogically, this ensures timely formative feedback for most classroom scenarios. A supplementary analysis discussing current security limitations and the evolving nature of security threats has been added, along with potential development ideas for enhancing system security. These improvements aim to strengthen the comprehensiveness and scientific reliability of our manuscript. Statistics show that 27.65 % of the students who participated in the evaluation were very satisfied with the feedback, while 17.4 % thought that the feedback was helpful. As for understanding the assessment content, 3.15 % of the students indicated that they needed more clarity, indicating that the clarity of the assessment questions still required improvement. Among the students’ learning performance, 67.24 % scored higher than the passing line, indicating that most students can master the course content.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"9 ","pages":"Article 100515"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145681344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predictive capability of foundational concepts tests for problem-solving using machine learning concepts: Evaluating project-based learning courses in artificial intelligence literacy education 使用机器学习概念解决问题的基础概念测试的预测能力:评估人工智能素养教育中基于项目的学习课程
Q1 Social Sciences Pub Date : 2025-12-01 DOI: 10.1016/j.caeai.2025.100503
Siu Cheung Kong , Chunyu Hou
In the artificial intelligence (AI) era, secondary and university students should be able to apply AI for problem-solving. This study designed and evaluated an AI literacy programme to enhance understanding of machine learning concepts. It also examined how the conceptual understanding from two foundational courses (Courses 1 and 2) affected students' application of these concepts in the subsequent two project-based learning courses (Courses 3 and 4). The regression analysis of data from 566, 566, 470, and 196 student participants enrolled on Courses 1, 2, 3, and 4, respectively, revealed that the post-course concept tests for Courses 1 and 2 accounted for 19.9 % of the variance in the students' problem-solving ability test before they took Course 3. This result indicates that teaching students' foundational concepts can develop their ability to solve machine learning-related problems. The post-course concept tests for Courses 1 and 2, together with the pre-course problem-solving ability test for Course 3, collectively explained 27.4 % of the variance in the students’ problem-solving ability after completing Course 3. Together with the significant improvement in the paired-samples t-test statistics for the pre- and post-course problem-solving test of Course 3, this indicates the importance of providing opportunities for students to solve machine learning-related problems. These findings provide empirical evidence to inform the design of curricula for AI literacy programmes. Project-based learning (PBL) is an approach that can provide opportunities for participants to develop problem-solving skills using foundational AI knowledge.
在人工智能(AI)时代,中学生和大学生应该能够应用AI解决问题。本研究设计并评估了一个人工智能扫盲计划,以提高对机器学习概念的理解。它还研究了两个基础课程(课程1和2)的概念理解如何影响学生在随后的两个基于项目的学习课程(课程3和4)中对这些概念的应用。通过对566名、566名、470名和196名分别参加课程1、2、3和4的学生进行回归分析,发现课程1和课程2的课后概念测试占课程3前学生问题解决能力测试方差的19.9%。这一结果表明,教授学生基本概念可以培养他们解决机器学习相关问题的能力。课程1和课程2的课程后概念测试和课程3的课程前问题解决能力测试共同解释了学生完成课程3后问题解决能力差异的27.4%。再加上课程3的课前和课后问题解决测试配对样本t检验统计量的显著提高,这表明为学生提供解决机器学习相关问题的机会的重要性。这些发现为人工智能扫盲课程的设计提供了经验证据。基于项目的学习(PBL)是一种方法,可以为参与者提供机会,发展使用基础人工智能知识解决问题的技能。
{"title":"Predictive capability of foundational concepts tests for problem-solving using machine learning concepts: Evaluating project-based learning courses in artificial intelligence literacy education","authors":"Siu Cheung Kong ,&nbsp;Chunyu Hou","doi":"10.1016/j.caeai.2025.100503","DOIUrl":"10.1016/j.caeai.2025.100503","url":null,"abstract":"<div><div>In the artificial intelligence (AI) era, secondary and university students should be able to apply AI for problem-solving. This study designed and evaluated an AI literacy programme to enhance understanding of machine learning concepts. It also examined how the conceptual understanding from two foundational courses (Courses 1 and 2) affected students' application of these concepts in the subsequent two project-based learning courses (Courses 3 and 4). The regression analysis of data from 566, 566, 470, and 196 student participants enrolled on Courses 1, 2, 3, and 4, respectively, revealed that the post-course concept tests for Courses 1 and 2 accounted for 19.9 % of the variance in the students' problem-solving ability test before they took Course 3. This result indicates that teaching students' foundational concepts can develop their ability to solve machine learning-related problems. The post-course concept tests for Courses 1 and 2, together with the pre-course problem-solving ability test for Course 3, collectively explained 27.4 % of the variance in the students’ problem-solving ability after completing Course 3. Together with the significant improvement in the paired-samples <em>t</em>-test statistics for the pre- and post-course problem-solving test of Course 3, this indicates the importance of providing opportunities for students to solve machine learning-related problems. These findings provide empirical evidence to inform the design of curricula for AI literacy programmes. Project-based learning (PBL) is an approach that can provide opportunities for participants to develop problem-solving skills using foundational AI knowledge.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"9 ","pages":"Article 100503"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145614501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards responsible AI in education: Challenges and implications for research and practice 在教育中实现负责任的人工智能:对研究和实践的挑战和影响
Q1 Social Sciences Pub Date : 2025-12-01 DOI: 10.1016/j.caeai.2024.100345
Teresa Cerratto Pargman, Cormac McGrath, Marcelo Milrad
{"title":"Towards responsible AI in education: Challenges and implications for research and practice","authors":"Teresa Cerratto Pargman,&nbsp;Cormac McGrath,&nbsp;Marcelo Milrad","doi":"10.1016/j.caeai.2024.100345","DOIUrl":"10.1016/j.caeai.2024.100345","url":null,"abstract":"","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"9 ","pages":"Article 100345"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145789991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing teachers’ AI competency: A professional development intervention study based on intelligent-TPACK framework 提升教师人工智能能力:基于智能- tpack框架的专业发展干预研究
Q1 Social Sciences Pub Date : 2025-12-01 DOI: 10.1016/j.caeai.2025.100521
Xiao Tan, Gary Cheng, Man Ho Ling
With the rapid penetration of generative artificial intelligence (AI) in higher education, university teachers' AI competency has become a critical determinant of effective technology integration in teaching. However, systematic and empirically validated intervention frameworks to support the development of this competency remain scarce. To address this gap, this study implemented a six-month professional development (PD) programme grounded in the Intelligent-TPACK framework and evaluated its effectiveness using a quasi-experimental pre-test-post-test design. A total of 64 teachers participated in the PD programme (experimental group), while pre- and post-test data were also collected from 61 teachers who did not participate (control group). Results indicate that the PD programme significantly enhanced AI competency in the experimental group, particularly in the domains of AI Technological Knowledge (AITK) and AI Technological Pedagogical Knowledge (AITPK). After controlling for baseline differences using ANCOVA, the effect size remained above the moderate threshold. A mixed-designed ANOVA further confirmed a significant interaction effect between group and time, ruling out maturation effects. Multi-level regression analysis revealed that background variables such as teaching experience, discipline, and professional title had limited predictive power for AI competency gains. Notably, self-perceived participation level did not significantly predict outcomes, whereas attendance rate emerged as a significant positive predictor. Interestingly, negative gain scores were observed in both groups. Follow-up interviews indicated that these scores did not reflect an actual decline in AI competency but rather a metacognitive recalibration, in which teachers shifted from unconscious incompetence to conscious incompetence—a pattern consistent with the Dunning–Kruger effect. This finding offers a novel theoretical perspective on the mechanism of change underlying the intervention. Overall, the PD programme based on the Intelligent-TPACK framework effectively enhanced university teachers’ AI competency and provides a systematic and evidence-based model for future PD initiatives in the AI era.
随着生成式人工智能(AI)在高等教育中的快速渗透,大学教师的AI能力已成为有效整合教学技术的关键决定因素。然而,支持这一能力发展的系统和经验验证的干预框架仍然很少。为了解决这一差距,本研究实施了一个基于智能- tpack框架的为期六个月的专业发展(PD)计划,并使用准实验的前测试-后测试设计评估其有效性。共有64名教师参加了PD项目(实验组),同时收集了61名未参加PD项目的教师(对照组)的测试前和测试后数据。结果表明,PD计划显著提高了实验组的人工智能能力,特别是在人工智能技术知识(AITK)和人工智能技术教学知识(AITPK)领域。在使用ANCOVA控制基线差异后,效应量仍高于中等阈值。混合设计的方差分析进一步证实了群体和时间之间的显著交互作用,排除了成熟效应。多层次回归分析显示,教学经验、学科、职称等背景变量对人工智能能力的预测能力有限。值得注意的是,自我感知的参与水平对结果没有显著的预测作用,而出勤率则是显著的正向预测因子。有趣的是,在两组中都观察到负增益分数。后续采访表明,这些分数并没有反映出人工智能能力的实际下降,而是一种元认知的重新校准,教师从无意识的无能转变为有意识的无能——这与邓宁-克鲁格效应一致。这一发现为干预背后的变化机制提供了一个新的理论视角。总体而言,基于Intelligent-TPACK框架的PD计划有效地提高了大学教师的人工智能能力,并为人工智能时代未来的PD计划提供了系统和基于证据的模型。
{"title":"Enhancing teachers’ AI competency: A professional development intervention study based on intelligent-TPACK framework","authors":"Xiao Tan,&nbsp;Gary Cheng,&nbsp;Man Ho Ling","doi":"10.1016/j.caeai.2025.100521","DOIUrl":"10.1016/j.caeai.2025.100521","url":null,"abstract":"<div><div>With the rapid penetration of generative artificial intelligence (AI) in higher education, university teachers' AI competency has become a critical determinant of effective technology integration in teaching. However, systematic and empirically validated intervention frameworks to support the development of this competency remain scarce. To address this gap, this study implemented a six-month professional development (PD) programme grounded in the Intelligent-TPACK framework and evaluated its effectiveness using a quasi-experimental pre-test-post-test design. A total of 64 teachers participated in the PD programme (experimental group), while pre- and post-test data were also collected from 61 teachers who did not participate (control group). Results indicate that the PD programme significantly enhanced AI competency in the experimental group, particularly in the domains of AI Technological Knowledge (AITK) and AI Technological Pedagogical Knowledge (AITPK). After controlling for baseline differences using ANCOVA, the effect size remained above the moderate threshold. A mixed-designed ANOVA further confirmed a significant interaction effect between group and time, ruling out maturation effects. Multi-level regression analysis revealed that background variables such as teaching experience, discipline, and professional title had limited predictive power for AI competency gains. Notably, self-perceived participation level did not significantly predict outcomes, whereas attendance rate emerged as a significant positive predictor. Interestingly, negative gain scores were observed in both groups. Follow-up interviews indicated that these scores did not reflect an actual decline in AI competency but rather a metacognitive recalibration, in which teachers shifted from unconscious incompetence to conscious incompetence—a pattern consistent with the Dunning–Kruger effect. This finding offers a novel theoretical perspective on the mechanism of change underlying the intervention. Overall, the PD programme based on the Intelligent-TPACK framework effectively enhanced university teachers’ AI competency and provides a systematic and evidence-based model for future PD initiatives in the AI era.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"9 ","pages":"Article 100521"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145736669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The effectiveness of an AI-integrated VR oral training application in reducing public speaking anxiety and interview anxiety 人工智能集成虚拟现实口语训练应用在减少公众演讲焦虑和面试焦虑中的效果
Q1 Social Sciences Pub Date : 2025-11-29 DOI: 10.1016/j.caeai.2025.100514
Peiwen Huang , Yanling Hwang , Jui Ling Hsu , Chien Fand Peng , Cheng Han Tsai , Chih Yao Wang
Despite the growing importance of English oral communication skills, traditional language learning approaches show limited effectiveness in simultaneously addressing psychological barriers and speaking proficiency among college students. While previous studies have explored anxiety reduction or speaking enhancement separately, a significant gap exists in research examining integrated approaches that tackle Public Speaking Anxiety (PSA), Interview Anxiety, and English-speaking proficiency improvement simultaneously. This study investigated whether an AI-integrated VR oral training application could effectively address these interconnected challenges. A quasi-experimental design was employed with 20 English major students from a mid-central university in Taiwan. Participants completed five training sessions using Meta Quest 2 headsets and an AI-integrated VR oral training application providing tailored feedback on pronunciation, grammar, and fluency based on IELTS standards. Pre- and post-intervention assessments utilized validated instruments including the Personal Report of Public Speaking Anxiety (PRPSA) and Measure of Anxiety in Selection Interviews (MASI), alongside comprehensive speaking proficiency measures. Results demonstrated significant improvements in English speaking proficiency, including increased sentence length and word count, with grammatical errors and incomplete sentences decreasing markedly (p < .001). Concurrently, significant reductions in both PRPSA and MASI scores (p < .05) were observed, though lexical diversity showed slight decline. VR-related motion-sickness symptoms were mildly alleviated, and participants' perceived control increased significantly (p < .05), while interest and attention levels remained stable. These findings suggest that AI-integrated VR oral training applications can effectively enhance English speaking proficiency while simultaneously reducing anxiety levels and improving self-efficacy among English learners. The study addresses a critical research gap by demonstrating the potential of integrated technological approaches to tackle multiple barriers to effective English oral communication, offering promising implications for language education and anxiety management in academic contexts.
尽管英语口语交际能力的重要性日益增加,但传统的语言学习方法在同时解决大学生的心理障碍和口语能力方面的效果有限。虽然以前的研究分别探讨了减少焦虑或提高口语能力,但同时研究解决公共演讲焦虑(PSA)、面试焦虑和提高英语水平的综合方法的研究存在显著差距。本研究调查了人工智能集成的VR口语训练应用程序是否可以有效地解决这些相互关联的挑战。本研究采用准实验设计,以20名台湾中部某大学英语专业学生为研究对象。参与者使用Meta Quest 2耳机和人工智能集成的VR口语训练应用程序完成了五个培训课程,该应用程序根据雅思标准提供量身定制的发音、语法和流利度反馈。干预前和干预后的评估使用了有效的工具,包括个人演讲焦虑报告(PRPSA)和选择访谈焦虑测量(MASI),以及综合口语能力测量。结果显示,英语口语水平显著提高,包括句子长度和字数增加,语法错误和不完整句子显著减少(p < .001)。同时,PRPSA和MASI得分显著下降(p < 0.05),尽管词汇多样性略有下降。vr相关的晕动病症状轻度缓解,参与者感知控制显著增加(p < 0.05),而兴趣和注意力水平保持稳定。这些研究结果表明,人工智能集成VR口语训练应用可以有效提高英语口语水平,同时降低英语学习者的焦虑水平,提高自我效能感。该研究通过展示综合技术方法解决有效英语口语交流的多重障碍的潜力,填补了一个关键的研究空白,为学术环境中的语言教育和焦虑管理提供了有希望的启示。
{"title":"The effectiveness of an AI-integrated VR oral training application in reducing public speaking anxiety and interview anxiety","authors":"Peiwen Huang ,&nbsp;Yanling Hwang ,&nbsp;Jui Ling Hsu ,&nbsp;Chien Fand Peng ,&nbsp;Cheng Han Tsai ,&nbsp;Chih Yao Wang","doi":"10.1016/j.caeai.2025.100514","DOIUrl":"10.1016/j.caeai.2025.100514","url":null,"abstract":"<div><div>Despite the growing importance of English oral communication skills, traditional language learning approaches show limited effectiveness in simultaneously addressing psychological barriers and speaking proficiency among college students. While previous studies have explored anxiety reduction or speaking enhancement separately, a significant gap exists in research examining integrated approaches that tackle Public Speaking Anxiety (PSA), Interview Anxiety, and English-speaking proficiency improvement simultaneously. This study investigated whether an AI-integrated VR oral training application could effectively address these interconnected challenges. A quasi-experimental design was employed with 20 English major students from a mid-central university in Taiwan. Participants completed five training sessions using Meta Quest 2 headsets and an AI-integrated VR oral training application providing tailored feedback on pronunciation, grammar, and fluency based on IELTS standards. Pre- and post-intervention assessments utilized validated instruments including the Personal Report of Public Speaking Anxiety (PRPSA) and Measure of Anxiety in Selection Interviews (MASI), alongside comprehensive speaking proficiency measures. Results demonstrated significant improvements in English speaking proficiency, including increased sentence length and word count, with grammatical errors and incomplete sentences decreasing markedly (p &lt; .001). Concurrently, significant reductions in both PRPSA and MASI scores (p &lt; .05) were observed, though lexical diversity showed slight decline. VR-related motion-sickness symptoms were mildly alleviated, and participants' perceived control increased significantly (p &lt; .05), while interest and attention levels remained stable. These findings suggest that AI-integrated VR oral training applications can effectively enhance English speaking proficiency while simultaneously reducing anxiety levels and improving self-efficacy among English learners. The study addresses a critical research gap by demonstrating the potential of integrated technological approaches to tackle multiple barriers to effective English oral communication, offering promising implications for language education and anxiety management in academic contexts.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"10 ","pages":"Article 100514"},"PeriodicalIF":0.0,"publicationDate":"2025-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145926802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stretching AI's reach: Assessing an AI-driven feedback system for extended academic writing 扩展AI的范围:评估AI驱动的反馈系统用于扩展学术写作
Q1 Social Sciences Pub Date : 2025-11-28 DOI: 10.1016/j.caeai.2025.100511
Jim Lo , Christy Wong , Agnes Ng , Pinna Wong , Denise Cheung , Pauli Lai
Advances in large language models (LLMs) enable timely and scalable writing evaluation. Previous research has shown that LLM-driven conversational systems, such as ChatGPT, can provide feedback on short essays. However, it is unclear whether AI can effectively evaluate more demanding genres. This study investigates a custom-built writing feedback system developed at a Hong Kong university that uses OpenAI's GPT-4 Turbo (0125-preview) to provide rubric-based feedback on a 1500-word academic report. Guided by a detailed, rubric-aligned prompt, the system generated 333 feedback items from 37 undergraduates, which were analysed for accuracy, tone, and inclusion of examples. The analysis showed that most feedback was accurate and addressed both strengths and weaknesses, but over half lacked concrete examples. Often recycling phrases from rubric descriptors, the feedback was largely generic and occasionally inaccurate. Interview data from six students revealed that the AI feedback was valued for its coverage, efficiency, and constructive tone, yet its generic nature undermined its usefulness. Despite these limitations, students expressed interest in receiving both AI and teacher feedback for the efficiency and coverage that AI offers, alongside the specificity and relevance of teacher input. These findings suggest that employing a well-crafted prompt on an AI model with a large context window does not necessarily guarantee substantive feedback. Therefore, educators using AI-driven feedback systems should thoroughly assess these systems' capacity to handle extended academic writing. Future research could explore ways to refine prompts and system design for long-form writing assignments.
大型语言模型(llm)的进步使及时和可扩展的写作评估成为可能。先前的研究表明,法学硕士驱动的会话系统,如ChatGPT,可以为短文提供反馈。然而,人工智能是否能够有效地评估要求更高的游戏类型尚不清楚。本研究调查了一所香港大学开发的定制写作反馈系统,该系统使用OpenAI的GPT-4 Turbo (0125-preview)为1500字的学术报告提供基于规则的反馈。在一个详细的、规则一致的提示的指导下,该系统从37名本科生中生成了333个反馈项目,并对其准确性、语气和示例的包含情况进行了分析。分析表明,大多数反馈都是准确的,并指出了优点和缺点,但超过一半的反馈缺乏具体的例子。这些反馈经常重复使用标题描述符中的短语,大部分是通用的,偶尔也不准确。来自六名学生的采访数据显示,人工智能反馈因其覆盖面、效率和建设性的语气而受到重视,但其普遍性削弱了其实用性。尽管存在这些限制,但学生们表示有兴趣接受人工智能和教师对人工智能提供的效率和覆盖范围的反馈,以及教师输入的特殊性和相关性。这些发现表明,在具有大上下文窗口的AI模型上使用精心设计的提示不一定能保证实质性的反馈。因此,使用人工智能驱动的反馈系统的教育工作者应该彻底评估这些系统处理扩展学术写作的能力。未来的研究可能会探索改进长篇写作作业提示和系统设计的方法。
{"title":"Stretching AI's reach: Assessing an AI-driven feedback system for extended academic writing","authors":"Jim Lo ,&nbsp;Christy Wong ,&nbsp;Agnes Ng ,&nbsp;Pinna Wong ,&nbsp;Denise Cheung ,&nbsp;Pauli Lai","doi":"10.1016/j.caeai.2025.100511","DOIUrl":"10.1016/j.caeai.2025.100511","url":null,"abstract":"<div><div>Advances in large language models (LLMs) enable timely and scalable writing evaluation. Previous research has shown that LLM-driven conversational systems, such as ChatGPT, can provide feedback on short essays. However, it is unclear whether AI can effectively evaluate more demanding genres. This study investigates a custom-built writing feedback system developed at a Hong Kong university that uses OpenAI's GPT-4 Turbo (0125-preview) to provide rubric-based feedback on a 1500-word academic report. Guided by a detailed, rubric-aligned prompt, the system generated 333 feedback items from 37 undergraduates, which were analysed for accuracy, tone, and inclusion of examples. The analysis showed that most feedback was accurate and addressed both strengths and weaknesses, but over half lacked concrete examples. Often recycling phrases from rubric descriptors, the feedback was largely generic and occasionally inaccurate. Interview data from six students revealed that the AI feedback was valued for its coverage, efficiency, and constructive tone, yet its generic nature undermined its usefulness. Despite these limitations, students expressed interest in receiving both AI and teacher feedback for the efficiency and coverage that AI offers, alongside the specificity and relevance of teacher input. These findings suggest that employing a well-crafted prompt on an AI model with a large context window does not necessarily guarantee substantive feedback. Therefore, educators using AI-driven feedback systems should thoroughly assess these systems' capacity to handle extended academic writing. Future research could explore ways to refine prompts and system design for long-form writing assignments.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"10 ","pages":"Article 100511"},"PeriodicalIF":0.0,"publicationDate":"2025-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145718922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computers and Education Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1