首页 > 最新文献

Assessment in Education-Principles Policy & Practice最新文献

英文 中文
Evaluating validity and bias for hand-calculated and automated written expression curriculum-based measurement scores 评估手工计算和自动书面表达课程测量分数的效度和偏倚
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2022-02-28 DOI: 10.1080/0969594X.2022.2043240
M. Matta, Sterett H. Mercer, Milena A. Keller-Margulis
ABSTRACT Written expression curriculum-based measurement (WE-CBM) is a formative assessment approach for screening and progress monitoring. To extend evaluation of WE-CBM, we compared hand-calculated and automated scoring approaches in relation to the number of screening samples needed per student for valid scores, the long-term predictive validity and diagnostic accuracy of scores, and predictive and diagnostic bias for underrepresented student groups. Second- to fifth-grade students (n = 609) completed five WE-CBM tasks during one academic year and a standardised writing test in fourth and seventh grade. Averaging WE-CBM scores across multiple samples improved validity. Complex hand-calculated metrics and automated tools outperformed simpler metrics for the long-term prediction of writing performance. No evidence of bias was observed between African American and Hispanic students. The study will illustrate the absence of test bias as necessary condition for fair and equitable screening procedures and the importance of future research to include comparisons with majority groups.
基于课程的书面表达测量(WE-CBM)是一种筛选和进度监测的形成性评估方法。为了扩展we - cbm的评估,我们比较了手动计算和自动评分方法与每个学生获得有效分数所需的筛选样本数量、分数的长期预测有效性和诊断准确性,以及对代表性不足的学生群体的预测和诊断偏差。二年级到五年级的学生(n = 609)在一个学年完成了五个WE-CBM任务,并在四年级和七年级完成了一个标准化写作测试。在多个样本中平均WE-CBM分数提高了效度。复杂的手工计算指标和自动化工具在长期预测写作性能方面优于简单的指标。没有证据表明在非裔美国人和西班牙裔学生之间存在偏见。该研究将说明测试偏差的缺失是公平和公平筛选程序的必要条件,以及未来研究包括与大多数群体比较的重要性。
{"title":"Evaluating validity and bias for hand-calculated and automated written expression curriculum-based measurement scores","authors":"M. Matta, Sterett H. Mercer, Milena A. Keller-Margulis","doi":"10.1080/0969594X.2022.2043240","DOIUrl":"https://doi.org/10.1080/0969594X.2022.2043240","url":null,"abstract":"ABSTRACT Written expression curriculum-based measurement (WE-CBM) is a formative assessment approach for screening and progress monitoring. To extend evaluation of WE-CBM, we compared hand-calculated and automated scoring approaches in relation to the number of screening samples needed per student for valid scores, the long-term predictive validity and diagnostic accuracy of scores, and predictive and diagnostic bias for underrepresented student groups. Second- to fifth-grade students (n = 609) completed five WE-CBM tasks during one academic year and a standardised writing test in fourth and seventh grade. Averaging WE-CBM scores across multiple samples improved validity. Complex hand-calculated metrics and automated tools outperformed simpler metrics for the long-term prediction of writing performance. No evidence of bias was observed between African American and Hispanic students. The study will illustrate the absence of test bias as necessary condition for fair and equitable screening procedures and the importance of future research to include comparisons with majority groups.","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"1 1","pages":"200 - 218"},"PeriodicalIF":3.2,"publicationDate":"2022-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74796485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Investigating the promise of automated writing evaluation for supporting formative writing assessment at scale 调查自动化写作评估的前景,以支持大规模的形成性写作评估
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2022-01-23 DOI: 10.1080/0969594X.2022.2025762
Joshua Wilson, Matthew C. Myers, Andrew Potter
ABSTRACT We investigated the promise of a novel approach to formative writing assessment at scale that involved an automated writing evaluation (AWE) system called MI Write. Specifically, we investigated elementary teachers’ perceptions and implementation of MI Write and changes in students’ writing performance in three genres from Fall to Spring associated with this implementation. Teachers in Grades 3–5 (n = 14) reported that MI Write was usable and acceptable, useful, and desirable; however, teachers tended to implement MI Write in a limited manner. Multilevel repeated measures analyses indicated that students in Grades 3–5 (n = 570) tended not to increase their performance from Fall to Spring except for third graders in all genres and fourth graders’ narrative writing. Findings illustrate the importance of educators utilising scalable formative assessments to evaluate and adjust core instruction.
我们研究了一种大规模形成性写作评估的新方法,该方法涉及一种名为MI Write的自动写作评估(AWE)系统。具体而言,我们调查了小学教师对MI写作的认知和实施,以及学生在秋季到春季三种类型的写作表现的变化。3-5年级的教师(n = 14)报告说MI Write是可用的、可接受的、有用的和可取的;然而,教师倾向于以有限的方式实施MI写作。多水平重复测量分析表明,除了三年级所有体裁和四年级叙事写作外,3-5年级学生(n = 570)从秋季到春季的成绩没有增加的趋势。研究结果说明了教育工作者利用可扩展的形成性评估来评估和调整核心教学的重要性。
{"title":"Investigating the promise of automated writing evaluation for supporting formative writing assessment at scale","authors":"Joshua Wilson, Matthew C. Myers, Andrew Potter","doi":"10.1080/0969594X.2022.2025762","DOIUrl":"https://doi.org/10.1080/0969594X.2022.2025762","url":null,"abstract":"ABSTRACT We investigated the promise of a novel approach to formative writing assessment at scale that involved an automated writing evaluation (AWE) system called MI Write. Specifically, we investigated elementary teachers’ perceptions and implementation of MI Write and changes in students’ writing performance in three genres from Fall to Spring associated with this implementation. Teachers in Grades 3–5 (n = 14) reported that MI Write was usable and acceptable, useful, and desirable; however, teachers tended to implement MI Write in a limited manner. Multilevel repeated measures analyses indicated that students in Grades 3–5 (n = 570) tended not to increase their performance from Fall to Spring except for third graders in all genres and fourth graders’ narrative writing. Findings illustrate the importance of educators utilising scalable formative assessments to evaluate and adjust core instruction.","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"82 1","pages":"183 - 199"},"PeriodicalIF":3.2,"publicationDate":"2022-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72999246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Assessment and learning in times of change and uncertainty 在变化和不确定的时代进行评估和学习
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2022-01-02 DOI: 10.1080/0969594X.2022.2068480
Therese N. Hopfenbeck
The year we left behind, 2021, was a year like no other, with the pandemic impacting students’ learning and assessment globally. While some countries have decided to open up, other countries have adopted a stricter approach and still use lockdown in areas of COVID outbreak. The impact of school lockdowns for students’ learning and the assessment of their achievements following pandemic will be the theme for a forthcoming Special Issue in this journal (O’Leary & Hayward to be published). What we already know, is that we are still in challenging times, where uncertainty is in the centre of our lives. Even worse, the uncertainty now includes dreadful scenes from the war in Ukraine, where children, instead of going to school and preparing for their exams this Spring, are fleeing their country in fear of bombs. Wars, whether they occur in Afghanistan, Ethiopia, Syria, or Yemen, are taking away the future of a generation of children who should be living together peacefully, connecting through music and art, preparing for adulthood, and developing the skills to solve problems we have not yet discovered. In challenging times, with a global pandemic, wars and personal losses, it is worth reminding ourselves of the reasons so many of us committed our lives to education. In Europe, the President of the assessment organisation AEA-Europe, Dr Christina Wikström, made the following statement to members in March 2022: ‘We place our trust in all our members, irrespective of nationality, to always stand up for peace, democracy, and equality for all. We must also work together in our scholarly profession since scholarship is necessary for the survival and prosperity of humanity’ (Wikström, 2022). It is in this spirit we publish the first 2022 issue of Assessment in Education – knowing that education matters, and particularly in times of uncertainty and challenges. As an international journal, we hope that researchers will continue to connect, collaborate, and stand up for the values which enhance learning for all students, no matter which continent they live on. We owe it to them to offer aspirations for the future, and quality education for all can provide that hope. In the first article in this regular issue, Steinmann et al. (this issue) presents findings from a study investigating the reliability of questionnaire scales. Student self-report scales have been known for their limitations, both with respect to validity and reliability (Samuelstuen & Bråten, 2007; Samuelstuen et al., 2007), but the current study specifically investigates the use of mixed wording in items used in international large-scale studies. More specifically, students were asked to report their agreement on items, such as ‘I usually do well in mathematics’ and ‘I am just not good at mathematics’, mixing positive and negative statements. Steinmann et al. used data from the international IEA studies PIRLS and TIMSS 2011, to further investigate how students responded to these scales, and demonst
我们离开的那一年,即2021年,是与众不同的一年,疫情影响了全球学生的学习和评估。虽然一些国家决定开放,但其他国家采取了更严格的措施,仍在疫情暴发地区采取封锁措施。大流行后学校封锁对学生学习和成绩评估的影响将是本杂志即将出版的一期特刊的主题(O ' leary & Hayward即将出版)。我们已经知道的是,我们仍然处于充满挑战的时代,不确定性是我们生活的中心。更糟糕的是,现在的不确定性包括乌克兰战争中的可怕场景,那里的孩子们因为害怕炸弹而逃离了他们的国家,而不是去上学,准备今年春天的考试。战争,无论是发生在阿富汗、埃塞俄比亚、叙利亚还是也门,都在夺走这一代儿童的未来。他们本应和平共处,通过音乐和艺术相互联系,为成年做准备,并培养解决我们尚未发现的问题的技能。在全球大流行、战争和个人损失的充满挑战的时代,我们有必要提醒自己,我们这么多人献出生命致力于教育的原因。在欧洲,评估组织欧洲原子能机构(AEA-Europe)主席克里斯蒂娜博士Wikström于2022年3月向成员发表了以下声明:“我们相信所有成员,无论国籍,都将始终捍卫所有人的和平、民主和平等。”我们也必须在我们的学术职业中共同努力,因为学术是人类生存和繁荣的必要条件”(Wikström, 2022)。正是本着这种精神,我们出版了2022年第一期《教育评估》——我们深知教育至关重要,尤其是在充满不确定性和挑战的时代。作为一份国际期刊,我们希望研究人员将继续联系、合作,并为促进所有学生学习的价值观挺身而出,无论他们生活在哪个大洲。我们有责任为他们提供对未来的憧憬,而全民优质教育可以提供这种希望。在本期的第一篇文章中,Steinmann等人(本期)提出了一项调查问卷量表可靠性的研究结果。学生自我报告量表以其有效性和可靠性方面的局限性而闻名(Samuelstuen & bramatten, 2007;Samuelstuen et al., 2007),但目前的研究专门调查了国际大规模研究中使用的项目中混合措辞的使用情况。更具体地说,学生们被要求报告他们对一些问题的认同,比如“我通常数学学得很好”和“我只是不擅长数学”,其中包括积极和消极的陈述。Steinmann等人使用国际能源署研究PIRLS和TIMSS 2011的数据,进一步调查了学生对这些量表的反应,并证明了2-36%的评估在教育中:原则,政策和实践2022,VOL. 29, NO. 6。1,1 - 4 https://doi.org/10.1080/0969594X.2022.2068480
{"title":"Assessment and learning in times of change and uncertainty","authors":"Therese N. Hopfenbeck","doi":"10.1080/0969594X.2022.2068480","DOIUrl":"https://doi.org/10.1080/0969594X.2022.2068480","url":null,"abstract":"The year we left behind, 2021, was a year like no other, with the pandemic impacting students’ learning and assessment globally. While some countries have decided to open up, other countries have adopted a stricter approach and still use lockdown in areas of COVID outbreak. The impact of school lockdowns for students’ learning and the assessment of their achievements following pandemic will be the theme for a forthcoming Special Issue in this journal (O’Leary & Hayward to be published). What we already know, is that we are still in challenging times, where uncertainty is in the centre of our lives. Even worse, the uncertainty now includes dreadful scenes from the war in Ukraine, where children, instead of going to school and preparing for their exams this Spring, are fleeing their country in fear of bombs. Wars, whether they occur in Afghanistan, Ethiopia, Syria, or Yemen, are taking away the future of a generation of children who should be living together peacefully, connecting through music and art, preparing for adulthood, and developing the skills to solve problems we have not yet discovered. In challenging times, with a global pandemic, wars and personal losses, it is worth reminding ourselves of the reasons so many of us committed our lives to education. In Europe, the President of the assessment organisation AEA-Europe, Dr Christina Wikström, made the following statement to members in March 2022: ‘We place our trust in all our members, irrespective of nationality, to always stand up for peace, democracy, and equality for all. We must also work together in our scholarly profession since scholarship is necessary for the survival and prosperity of humanity’ (Wikström, 2022). It is in this spirit we publish the first 2022 issue of Assessment in Education – knowing that education matters, and particularly in times of uncertainty and challenges. As an international journal, we hope that researchers will continue to connect, collaborate, and stand up for the values which enhance learning for all students, no matter which continent they live on. We owe it to them to offer aspirations for the future, and quality education for all can provide that hope. In the first article in this regular issue, Steinmann et al. (this issue) presents findings from a study investigating the reliability of questionnaire scales. Student self-report scales have been known for their limitations, both with respect to validity and reliability (Samuelstuen & Bråten, 2007; Samuelstuen et al., 2007), but the current study specifically investigates the use of mixed wording in items used in international large-scale studies. More specifically, students were asked to report their agreement on items, such as ‘I usually do well in mathematics’ and ‘I am just not good at mathematics’, mixing positive and negative statements. Steinmann et al. used data from the international IEA studies PIRLS and TIMSS 2011, to further investigate how students responded to these scales, and demonst","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"51 1","pages":"1 - 4"},"PeriodicalIF":3.2,"publicationDate":"2022-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86513103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Should they change their answers or not? Modelling Achievement through a Metacognitive Lens 他们是否应该改变他们的答案?通过元认知镜头建模成就
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2022-01-02 DOI: 10.1080/0969594X.2022.2053945
E. Papanastasiou, Agni Stylianou-Georgiou
ABSTRACT Α frequently used indicator to reflect student performance is that of a test score. However, although tests are designed to assess students’ knowledge or skills, other factors can also affect test results such as test-taking strategies. Therefore, the purpose of this study was to model the interrelationships among test-taking strategy instruction (with a focus on metacognition), answer changing bias and performance on multiple-choice tests among college students through structural equation modelling. This study, which was based on a sample of 1512 students from Greece, has managed to extend the findings of previous research by proposing a model that demonstrates the interplay between these variables.
摘要Α反映学生表现的常用指标是考试成绩。然而,尽管考试是为了评估学生的知识或技能而设计的,但其他因素也会影响考试结果,比如考试策略。因此,本研究的目的是通过结构方程模型来模拟大学生应试策略指导(以元认知为重点)、答案改变偏差和多项选择题成绩之间的相互关系。这项研究以1512名希腊学生为样本,通过提出一个模型来展示这些变量之间的相互作用,成功地扩展了之前的研究结果。
{"title":"Should they change their answers or not? Modelling Achievement through a Metacognitive Lens","authors":"E. Papanastasiou, Agni Stylianou-Georgiou","doi":"10.1080/0969594X.2022.2053945","DOIUrl":"https://doi.org/10.1080/0969594X.2022.2053945","url":null,"abstract":"ABSTRACT Α frequently used indicator to reflect student performance is that of a test score. However, although tests are designed to assess students’ knowledge or skills, other factors can also affect test results such as test-taking strategies. Therefore, the purpose of this study was to model the interrelationships among test-taking strategy instruction (with a focus on metacognition), answer changing bias and performance on multiple-choice tests among college students through structural equation modelling. This study, which was based on a sample of 1512 students from Greece, has managed to extend the findings of previous research by proposing a model that demonstrates the interplay between these variables.","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"1 1","pages":"77 - 94"},"PeriodicalIF":3.2,"publicationDate":"2022-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82298116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Building useful, web-based educational assessment tools for students, with students: a demonstration with the school climate walkthrough 与学生一起为学生建立有用的、基于网络的教育评估工具:学校气候演练示范
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2022-01-02 DOI: 10.1080/0969594X.2022.2056576
Jessica D. Hoffmann, Rachel Baumsteiger, Jennifer Seibyl, E. Hills, Christina M. Bradley, Christina Cipriano, M. Brackett
ABSTRACT Practical educational assessment tools for adolescents champion student voice, use technology to enhance engagement, highlight discrepancies in school experience, and provide actionable feedback. We report a series of studies that detail an iterative process for developing a new school climate assessment tool: (1) item generation that centres student voice, (2) the design of a web-based app, (3) item revisions informed by student and educator feedback, and (4) the identification and confirmation of the underlying factor structure of the assessment tool. The resulting School Climate Walkthrough provides scores on nine dimensions of school climate and 73 additional observational items. The web-based application produces instantaneous reports that display systemic differences in how various student demographic groups experience school. This process can guide future research in building the next generation of educational assessments for adolescents and disrupting harmful or exclusionary school practices.
实用的青少年教育评估工具支持学生的声音,使用技术来提高参与度,突出学校体验的差异,并提供可操作的反馈。我们报告了一系列研究,详细介绍了开发新的学校气候评估工具的迭代过程:(1)以学生声音为中心的项目生成,(2)基于网络的应用程序的设计,(3)根据学生和教育者的反馈进行项目修改,以及(4)识别和确认评估工具的潜在因素结构。由此产生的学校气候演练提供了学校气候的九个维度和73个额外的观察项目的分数。这个基于网络的应用程序产生即时报告,显示不同学生群体在学校经历方面的系统差异。这一进程可以指导今后在建立下一代青少年教育评估和破坏有害或排他性学校做法方面的研究。
{"title":"Building useful, web-based educational assessment tools for students, with students: a demonstration with the school climate walkthrough","authors":"Jessica D. Hoffmann, Rachel Baumsteiger, Jennifer Seibyl, E. Hills, Christina M. Bradley, Christina Cipriano, M. Brackett","doi":"10.1080/0969594X.2022.2056576","DOIUrl":"https://doi.org/10.1080/0969594X.2022.2056576","url":null,"abstract":"ABSTRACT Practical educational assessment tools for adolescents champion student voice, use technology to enhance engagement, highlight discrepancies in school experience, and provide actionable feedback. We report a series of studies that detail an iterative process for developing a new school climate assessment tool: (1) item generation that centres student voice, (2) the design of a web-based app, (3) item revisions informed by student and educator feedback, and (4) the identification and confirmation of the underlying factor structure of the assessment tool. The resulting School Climate Walkthrough provides scores on nine dimensions of school climate and 73 additional observational items. The web-based application produces instantaneous reports that display systemic differences in how various student demographic groups experience school. This process can guide future research in building the next generation of educational assessments for adolescents and disrupting harmful or exclusionary school practices.","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"34 1","pages":"95 - 120"},"PeriodicalIF":3.2,"publicationDate":"2022-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82626616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Are high school students accurate in predicting their AP exam scores?: Examining inaccuracy and overconfidence of students’ predictions 高中生预测自己的AP考试成绩准确吗?:检查学生预测的不准确性和过度自信
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2022-01-02 DOI: 10.1080/0969594X.2022.2037508
Teresa M. Ober, Maxwell R. Hong, Matthew F. Carter, Alex S. Brodersen, Daniella A. Rebouças-Ju, Cheng Liu, Ying Cheng
ABSTRACT We examined whether students were accurate in predicting their test performance in both low-stakes and high-stakes testing contexts. The sample comprised U.S. high school students enrolled in an advanced placement (AP) statistics course during the 2017–2018 academic year (N = 209; Mage = 16.6 years). We found that even two months before taking the AP exam, a high stakes summative assessment, students were moderately accurate in predicting their actual scores (κweighted = .62). When the same variables were entered into models predicting inaccuracy and overconfidence bias, results did not provide evidence that age, gender, parental education, number of mathematics classes previously taken, or course engagement accounted for variation in accuracy. Overconfidence bias differed between students enrolled at different schools. Results indicated that students’ predictions of performance were positively associated with performance in both low- and high-stakes testing contexts. The findings shed light on ways to leverage students’ self-assessment for learning.
摘要:我们研究了学生在低风险和高风险测试情境下预测自己考试成绩的准确性。样本包括2017-2018学年参加AP统计学课程的美国高中生(N = 209;法师= 16.6年)。我们发现,即使在参加AP考试(一项高风险的总结性评估)前两个月,学生对自己实际分数的预测也有一定的准确性(κ加权= 0.62)。当同样的变量被输入到预测不准确性和过度自信偏差的模型中时,结果并没有提供证据表明年龄、性别、父母的教育程度、以前上过的数学课的数量或课程参与可以解释准确性的变化。不同学校的学生对过度自信的偏见有所不同。结果表明,在低风险和高风险的测试环境中,学生对成绩的预测与成绩呈正相关。这些发现揭示了如何利用学生的自我评估来学习。
{"title":"Are high school students accurate in predicting their AP exam scores?: Examining inaccuracy and overconfidence of students’ predictions","authors":"Teresa M. Ober, Maxwell R. Hong, Matthew F. Carter, Alex S. Brodersen, Daniella A. Rebouças-Ju, Cheng Liu, Ying Cheng","doi":"10.1080/0969594X.2022.2037508","DOIUrl":"https://doi.org/10.1080/0969594X.2022.2037508","url":null,"abstract":"ABSTRACT We examined whether students were accurate in predicting their test performance in both low-stakes and high-stakes testing contexts. The sample comprised U.S. high school students enrolled in an advanced placement (AP) statistics course during the 2017–2018 academic year (N = 209; Mage = 16.6 years). We found that even two months before taking the AP exam, a high stakes summative assessment, students were moderately accurate in predicting their actual scores (κweighted = .62). When the same variables were entered into models predicting inaccuracy and overconfidence bias, results did not provide evidence that age, gender, parental education, number of mathematics classes previously taken, or course engagement accounted for variation in accuracy. Overconfidence bias differed between students enrolled at different schools. Results indicated that students’ predictions of performance were positively associated with performance in both low- and high-stakes testing contexts. The findings shed light on ways to leverage students’ self-assessment for learning.","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"124 1","pages":"27 - 50"},"PeriodicalIF":3.2,"publicationDate":"2022-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79491546","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Thirty years of GCSE: A review of student views and experiences 三十年的GCSE:学生的观点和经验的回顾
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2022-01-02 DOI: 10.1080/0969594X.2022.2053946
K. Brown, Kevin Woods
ABSTRACT In England, Wales and Northern Ireland, the General Certificate of Secondary Education (GCSE) has been the qualification by which students’ attainment at age sixteen has been measured for the last thirty years. Despite the longevity of GCSEs, relatively little research has explored the views and experiences of those undertaking them. Using a systematic literature review methodology and critical appraisal frameworks, the current study synthesises the literature in this area in order to elucidate young people’s views and experiences of GCSE study and assessment. Findings suggest that although there are positive aspects of GCSE study and assessment, for some young people, GCSE study, assessment and recent reforms appear to be relatively negative experiences, characterised by low levels of enjoyment and well-being and high levels of stress and test anxiety. Findings also suggest that agency, equality and fairness and relatedness are important factors in mediating young people’s experiences of GCSE.
在英格兰、威尔士和北爱尔兰,普通中等教育证书(GCSE)在过去的30年里一直是衡量学生在16岁时的成就的资格证书。尽管普通中等教育证书的历史悠久,但相对较少的研究探讨了那些参加这些考试的人的观点和经历。采用系统的文献综述方法和批判性评估框架,本研究综合了这一领域的文献,以阐明年轻人对GCSE学习和评估的看法和经验。研究结果表明,虽然GCSE学习和评估有积极的方面,但对于一些年轻人来说,GCSE学习、评估和最近的改革似乎是相对消极的经历,其特点是低水平的享受和幸福,高水平的压力和考试焦虑。研究结果还表明,代理、平等、公平和相关性是中介青少年GCSE体验的重要因素。
{"title":"Thirty years of GCSE: A review of student views and experiences","authors":"K. Brown, Kevin Woods","doi":"10.1080/0969594X.2022.2053946","DOIUrl":"https://doi.org/10.1080/0969594X.2022.2053946","url":null,"abstract":"ABSTRACT In England, Wales and Northern Ireland, the General Certificate of Secondary Education (GCSE) has been the qualification by which students’ attainment at age sixteen has been measured for the last thirty years. Despite the longevity of GCSEs, relatively little research has explored the views and experiences of those undertaking them. Using a systematic literature review methodology and critical appraisal frameworks, the current study synthesises the literature in this area in order to elucidate young people’s views and experiences of GCSE study and assessment. Findings suggest that although there are positive aspects of GCSE study and assessment, for some young people, GCSE study, assessment and recent reforms appear to be relatively negative experiences, characterised by low levels of enjoyment and well-being and high levels of stress and test anxiety. Findings also suggest that agency, equality and fairness and relatedness are important factors in mediating young people’s experiences of GCSE.","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"348 1","pages":"51 - 76"},"PeriodicalIF":3.2,"publicationDate":"2022-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86798843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
The impact of inconsistent responders to mixed-worded scales on inferences in international large-scale assessments 国际大规模评估中混合用词量表不一致应答者对推论的影响
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-11-22 DOI: 10.1080/0969594X.2021.2005302
Isa Steinmann, Daniel Sánchez, Saskia van Laar, J. Braeken
ABSTRACT Questionnaire scales that are mixed-worded, i.e. include both positively and negatively worded items, often suffer from issues like low reliability and more complex latent structures than intended. Part of the problem might be that some responders fail to respond consistently to the mixed-worded items. We investigated the prevalence and impact of inconsistent responders in 37 primary education systems participating in the joint PIRLS/TIMSS 2011 assessment. Using the mean absolute difference method and three mixed-worded self-concept scales, we identified between 2%‒36% of students as inconsistent responders across education systems. Consistent with expectations, these students showed lower average achievement scores and had a higher risk of being identified as inconsistent on more than one scale. We also found that the inconsistent responders biased the estimated dimensionality and reliability of the scales. The impact on external validity measures was limited and unsystematic. We discuss implications for the use and development of questionnaire scales.
摘要问卷量表的混合措辞,即同时包含积极和消极措辞,往往存在可靠性低和潜在结构比预期更复杂等问题。这个问题的部分原因可能是一些响应者不能始终如一地对混合词的项目作出反应。我们调查了参与PIRLS/TIMSS 2011联合评估的37个小学教育系统中不一致响应者的患病率和影响。使用平均绝对差法和三个混合措辞的自我概念量表,我们确定了2%-36%的学生在不同的教育系统中是不一致的反应者。与预期一致的是,这些学生的平均成绩较低,而且在多个量表上被认定为不一致的风险更高。我们还发现不一致的应答者对量表的估计维度和可靠性有偏差。对外部效度测量的影响是有限的和不系统的。我们讨论了问卷量表的使用和发展的意义。
{"title":"The impact of inconsistent responders to mixed-worded scales on inferences in international large-scale assessments","authors":"Isa Steinmann, Daniel Sánchez, Saskia van Laar, J. Braeken","doi":"10.1080/0969594X.2021.2005302","DOIUrl":"https://doi.org/10.1080/0969594X.2021.2005302","url":null,"abstract":"ABSTRACT Questionnaire scales that are mixed-worded, i.e. include both positively and negatively worded items, often suffer from issues like low reliability and more complex latent structures than intended. Part of the problem might be that some responders fail to respond consistently to the mixed-worded items. We investigated the prevalence and impact of inconsistent responders in 37 primary education systems participating in the joint PIRLS/TIMSS 2011 assessment. Using the mean absolute difference method and three mixed-worded self-concept scales, we identified between 2%‒36% of students as inconsistent responders across education systems. Consistent with expectations, these students showed lower average achievement scores and had a higher risk of being identified as inconsistent on more than one scale. We also found that the inconsistent responders biased the estimated dimensionality and reliability of the scales. The impact on external validity measures was limited and unsystematic. We discuss implications for the use and development of questionnaire scales.","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"16 1","pages":"5 - 26"},"PeriodicalIF":3.2,"publicationDate":"2021-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83182542","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Assessment from a disciplinary approach: design and implementation in three undergraduate programmes 从学科方法评估:三个本科课程的设计和实施
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-11-02 DOI: 10.1080/0969594X.2021.1999210
J. Fernández-Ruiz, E. Panadero, Daniel García-Pérez
ABSTRACT The role of the academic discipline is a major factor in the assessment design and implementation in higher education. Unfortunately, a clear understanding of how teachers from different disciplines approach assessment is still missing; this information can lead to teacher training programmes that are better designed and more focussed. The present study compared assessment design and implementation in three programmes (sport science, mathematics, and medicine) each representing a discipline from 4 Spanish universities. Using a mixed-methods approach, data from syllabi (N = 385) and semi-structured interviews with teachers (N = 19) were analysed. The results showed different approaches to assessment design and implementation in each programme in two main axes: summative or formative purposes of assessment, and content-based or authentic assessment. Implications for further research are discussed.
学科的作用是影响高等教育评估设计与实施的重要因素。不幸的是,对于不同学科的教师如何进行评估,我们仍然缺乏清晰的认识;这些信息可以使教师培训方案设计得更好,重点更突出。本研究比较了三个项目(体育科学、数学和医学)的评估设计和实施,每个项目代表了来自4所西班牙大学的一个学科。采用混合方法,对来自教学大纲(N = 385)和教师半结构化访谈(N = 19)的数据进行分析。结果表明,在两个主要方面,每个方案的评估设计和执行方法不同:总结性或形成性评估目的,以及基于内容或真实的评估。讨论了进一步研究的意义。
{"title":"Assessment from a disciplinary approach: design and implementation in three undergraduate programmes","authors":"J. Fernández-Ruiz, E. Panadero, Daniel García-Pérez","doi":"10.1080/0969594X.2021.1999210","DOIUrl":"https://doi.org/10.1080/0969594X.2021.1999210","url":null,"abstract":"ABSTRACT The role of the academic discipline is a major factor in the assessment design and implementation in higher education. Unfortunately, a clear understanding of how teachers from different disciplines approach assessment is still missing; this information can lead to teacher training programmes that are better designed and more focussed. The present study compared assessment design and implementation in three programmes (sport science, mathematics, and medicine) each representing a discipline from 4 Spanish universities. Using a mixed-methods approach, data from syllabi (N = 385) and semi-structured interviews with teachers (N = 19) were analysed. The results showed different approaches to assessment design and implementation in each programme in two main axes: summative or formative purposes of assessment, and content-based or authentic assessment. Implications for further research are discussed.","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"18 4 1","pages":"703 - 723"},"PeriodicalIF":3.2,"publicationDate":"2021-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83603924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Formative assessment, growth mindset, and achievement: examining their relations in the East and the West 形成性评价、成长心态与成就:东西方关系考察
IF 3.2 3区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-10-31 DOI: 10.1080/0969594X.2021.1988510
Zi Yan, Ronnel B. King, Joseph Y. Haw
ABSTRACT Both formative assessment and growth mindset scholars aim to understand how to enhance achievement. While research on formative assessment focuses on external teaching practices, work on growth mindset emphasises internal psychological processes. This study examined the interplay between three formative assessment strategies (i.e. sharing learning progressions, providing feedback, and instructional adjustments) and growth mindset in predicting reading achievement using the PISA2018 data. We focused specifically on samples from the West (the United States, the United Kingdom, Ireland, Canada, Australia, and New Zealand) and the East (Mainland China, Hong Kong SAR, Macau SAR, Chinese Taipei, Japan and Korea) which comprised of 109,204 15-year old students. The results showed that formative assessment strategies were positively, albeit weakly, related to a growth mindset in the East, but not in the West. In contrast, growth mindset was positively related to reading achievement only in the West, but not in the East. The impacts of different formative assessment strategies on reading achievement demonstrated cross-cultural variability, but the strongest positive predictor was instructional adjustments. These findings highlight the potential synergy between formative assessment and growth mindset in enhancing academic achievement as well as the importance of cultural contexts in understanding their roles in student learning.
形成性评价和成长心态学者都致力于了解如何提高成就。形成性评价的研究侧重于外部教学实践,而成长性思维的研究则侧重于内部心理过程。本研究利用PISA2018数据研究了三种形成性评估策略(即分享学习进度、提供反馈和教学调整)与成长心态在预测阅读成绩方面的相互作用。我们特别关注来自西方(美国、英国、爱尔兰、加拿大、澳大利亚和新西兰)和东方(中国大陆、香港特别行政区、澳门特别行政区、中国台北、日本和韩国)的样本,其中包括109,204名15岁的学生。结果表明,形成性评估策略在东方与成长心态呈正相关,尽管关联度较弱,但在西方没有。相反,成长心态与阅读成绩的正相关仅在西方国家存在,而在东方没有。不同形成性评价策略对阅读成绩的影响表现出跨文化差异,但教学调整是最有效的正向预测因子。这些发现强调了形成性评价和成长性思维在提高学业成绩方面的潜在协同作用,以及文化背景在理解它们在学生学习中的作用方面的重要性。
{"title":"Formative assessment, growth mindset, and achievement: examining their relations in the East and the West","authors":"Zi Yan, Ronnel B. King, Joseph Y. Haw","doi":"10.1080/0969594X.2021.1988510","DOIUrl":"https://doi.org/10.1080/0969594X.2021.1988510","url":null,"abstract":"ABSTRACT Both formative assessment and growth mindset scholars aim to understand how to enhance achievement. While research on formative assessment focuses on external teaching practices, work on growth mindset emphasises internal psychological processes. This study examined the interplay between three formative assessment strategies (i.e. sharing learning progressions, providing feedback, and instructional adjustments) and growth mindset in predicting reading achievement using the PISA2018 data. We focused specifically on samples from the West (the United States, the United Kingdom, Ireland, Canada, Australia, and New Zealand) and the East (Mainland China, Hong Kong SAR, Macau SAR, Chinese Taipei, Japan and Korea) which comprised of 109,204 15-year old students. The results showed that formative assessment strategies were positively, albeit weakly, related to a growth mindset in the East, but not in the West. In contrast, growth mindset was positively related to reading achievement only in the West, but not in the East. The impacts of different formative assessment strategies on reading achievement demonstrated cross-cultural variability, but the strongest positive predictor was instructional adjustments. These findings highlight the potential synergy between formative assessment and growth mindset in enhancing academic achievement as well as the importance of cultural contexts in understanding their roles in student learning.","PeriodicalId":51515,"journal":{"name":"Assessment in Education-Principles Policy & Practice","volume":"16 1","pages":"676 - 702"},"PeriodicalIF":3.2,"publicationDate":"2021-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90676693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
Assessment in Education-Principles Policy & Practice
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1