首页 > 最新文献

Educational Assessment最新文献

英文 中文
Investigating the Effects of Test Accommodations with Process Data for English Learners in a Mathematics Assessment 用过程数据研究考试调整对英语学习者数学评估的影响
IF 1.5 Q2 Social Sciences Pub Date : 2021-09-29 DOI: 10.1080/10627197.2021.1982693
M. Wolf, Hanwook Yoo, Danielle Guzman-Orth, J. Abedi
ABSTRACT Implementing a randomized controlled trial design, the present study investigated the effects of two types of accommodations, linguistic modification and a glossary, for English learners (ELs) taking a computer-based mathematics assessment. Process data including response time and clicks on glossary words were also examined to better interpret students’ interaction with the accommodations in the testing conditions. Regression and ANOVA analyses were performed with data from 513 students (189 ELs and 324 non-ELs) in Grade 9. No statistically significant accommodation effects were detected in this study. Process data revealed possible explanations (i.e., student engagement and glossary usage) for the nonsignificant results. Implications for future research on test accommodations for EL students are discussed.
本研究采用随机对照试验设计,研究了语言修饰和词汇表两种辅助工具对英语学习者参加计算机数学评估的影响。为了更好地解释学生在测试条件下与住宿设施的互动,还检查了包括响应时间和对词汇表单词的点击在内的过程数据。对9年级513名学生(189名ei和324名非ei)的数据进行回归和方差分析。本研究未发现统计学上显著的调节效应。过程数据揭示了不显著结果的可能解释(即学生参与度和词汇使用)。本文还讨论了未来研究对外语学生考试便利的启示。
{"title":"Investigating the Effects of Test Accommodations with Process Data for English Learners in a Mathematics Assessment","authors":"M. Wolf, Hanwook Yoo, Danielle Guzman-Orth, J. Abedi","doi":"10.1080/10627197.2021.1982693","DOIUrl":"https://doi.org/10.1080/10627197.2021.1982693","url":null,"abstract":"ABSTRACT Implementing a randomized controlled trial design, the present study investigated the effects of two types of accommodations, linguistic modification and a glossary, for English learners (ELs) taking a computer-based mathematics assessment. Process data including response time and clicks on glossary words were also examined to better interpret students’ interaction with the accommodations in the testing conditions. Regression and ANOVA analyses were performed with data from 513 students (189 ELs and 324 non-ELs) in Grade 9. No statistically significant accommodation effects were detected in this study. Process data revealed possible explanations (i.e., student engagement and glossary usage) for the nonsignificant results. Implications for future research on test accommodations for EL students are discussed.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45298505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Methodology for Determining and Validating Latent Factor Dimensionality of Complex Multi-Factor Science Constructs Measuring Knowledge-In-Use 复杂多因素科学结构中潜在因素维度的确定与验证方法
IF 1.5 Q2 Social Sciences Pub Date : 2021-09-05 DOI: 10.1080/10627197.2021.1971966
Leonora Kaldaras, Hope O. Akaeze, J. Krajcik
ABSTRACT Deep science understanding is reflected in students’ ability to use content and skills when making sense of the world. Assessing deep understanding requires measuring complex constructs that combine elements of content and skills. To develop valid measures of complex constructs, we need to understand how their theoretical dimensionality, reflected in the integration of content and skills, is manifested in practice. This work is developed in the context of the Framework for K-12 Science Education and Next-Generation Science Standards (NGSS). We introduce a methodology that describes steps for creating a theoretical validity argument for measuring complex NGSS constructs, designing operational assessments based on this argument, and obtaining empirical evidence for the validity of the argument and assessments, focusing on how theoretically suggested dimensionality of NGSS constructs is manifested in practice. Results have implications for developing valid NGSS assessments and reporting student progress on high-stakes and diagnostic evaluation.
深刻的科学理解体现在学生在理解世界时运用内容和技能的能力上。评估深度理解需要测量结合内容和技能元素的复杂结构。为了开发复杂构念的有效测量方法,我们需要了解它们的理论维度(反映在内容和技能的整合中)是如何在实践中表现出来的。这项工作是在K-12科学教育框架和下一代科学标准(NGSS)的背景下开展的。我们介绍了一种方法,该方法描述了为测量复杂NGSS结构创建理论有效性论证的步骤,设计基于该论证的操作评估,以及获得论证和评估有效性的经验证据,重点关注理论建议的NGSS结构维度如何在实践中得到体现。研究结果对制定有效的NGSS评估和报告学生在高风险和诊断性评估方面的进展具有重要意义。
{"title":"A Methodology for Determining and Validating Latent Factor Dimensionality of Complex Multi-Factor Science Constructs Measuring Knowledge-In-Use","authors":"Leonora Kaldaras, Hope O. Akaeze, J. Krajcik","doi":"10.1080/10627197.2021.1971966","DOIUrl":"https://doi.org/10.1080/10627197.2021.1971966","url":null,"abstract":"ABSTRACT Deep science understanding is reflected in students’ ability to use content and skills when making sense of the world. Assessing deep understanding requires measuring complex constructs that combine elements of content and skills. To develop valid measures of complex constructs, we need to understand how their theoretical dimensionality, reflected in the integration of content and skills, is manifested in practice. This work is developed in the context of the Framework for K-12 Science Education and Next-Generation Science Standards (NGSS). We introduce a methodology that describes steps for creating a theoretical validity argument for measuring complex NGSS constructs, designing operational assessments based on this argument, and obtaining empirical evidence for the validity of the argument and assessments, focusing on how theoretically suggested dimensionality of NGSS constructs is manifested in practice. Results have implications for developing valid NGSS assessments and reporting student progress on high-stakes and diagnostic evaluation.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45735723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Assessing Source Evaluation Skills of Middle School Students Using Learning Progressions 利用学习进展评估中学生源评价技能
IF 1.5 Q2 Social Sciences Pub Date : 2021-09-01 DOI: 10.1080/10627197.2021.1966299
Jesse R. Sparks, P. V. van Rijn, P. Deane
ABSTRACT Effectively evaluating the credibility and accuracy of multiple sources is critical for college readiness. We developed 24 source evaluation tasks spanning four predicted difficulty levels of a hypothesized learning progression (LP) and piloted these tasks to evaluate the utility of an LP-based approach to designing formative literacy assessments. Sixth, seventh, and eighth grade students (N = 360, 120 per grade) completed 12 of the 24 tasks in an online testing session. Analyses examined the tasks’ reliability and validity and whether patterns of performance aligned to predicted LP levels (i.e., recovery of the LP) using task progression maps derived from item response theory (IRT). Results suggested that the LP tasks were reliable and correlated with external measures; however, some lower level tasks proved unexpectedly difficult. Possible explanations for low performance are discussed, followed by implications for future LP and task revisions. This work provides a model for designing and evaluating LP-based literacy assessments.
摘要有效评估多种来源的可信度和准确性对大学入学准备至关重要。我们开发了24个来源评估任务,涵盖假设学习进展(LP)的四个预测难度水平,并对这些任务进行了试点,以评估基于LP的方法在设计形成性识字评估中的效用。六年级、七年级和八年级的学生(每个年级360120名)在一次在线测试中完成了24项任务中的12项。分析使用从项目反应理论(IRT)得出的任务进展图检查了任务的可靠性和有效性,以及表现模式是否与预测的LP水平一致(即LP的恢复)。结果表明,LP任务是可靠的,并且与外部测量相关;然而,一些较低级别的任务却出乎意料地困难。讨论了低性能的可能解释,然后讨论了对未来LP和任务修订的影响。这项工作为设计和评估基于LP的识字评估提供了一个模型。
{"title":"Assessing Source Evaluation Skills of Middle School Students Using Learning Progressions","authors":"Jesse R. Sparks, P. V. van Rijn, P. Deane","doi":"10.1080/10627197.2021.1966299","DOIUrl":"https://doi.org/10.1080/10627197.2021.1966299","url":null,"abstract":"ABSTRACT Effectively evaluating the credibility and accuracy of multiple sources is critical for college readiness. We developed 24 source evaluation tasks spanning four predicted difficulty levels of a hypothesized learning progression (LP) and piloted these tasks to evaluate the utility of an LP-based approach to designing formative literacy assessments. Sixth, seventh, and eighth grade students (N = 360, 120 per grade) completed 12 of the 24 tasks in an online testing session. Analyses examined the tasks’ reliability and validity and whether patterns of performance aligned to predicted LP levels (i.e., recovery of the LP) using task progression maps derived from item response theory (IRT). Results suggested that the LP tasks were reliable and correlated with external measures; however, some lower level tasks proved unexpectedly difficult. Possible explanations for low performance are discussed, followed by implications for future LP and task revisions. This work provides a model for designing and evaluating LP-based literacy assessments.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41628487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
An Intersectional Approach to DIF: Do Initial Findings Hold across Tests? DIF的跨部门方法:初始发现在测试中有效吗?
IF 1.5 Q2 Social Sciences Pub Date : 2021-08-22 DOI: 10.1080/10627197.2021.1965473
M. Russell, Olivia Szendey, Larry Kaplan
ABSTRACT Differential Item Function (DIF) analysis is commonly employed to examine potential bias produced by a test item. Since its introduction DIF analyses have focused on potential bias related to broad categories of oppression, including gender, racial stratification, economic class, and ableness. More recently, efforts to examine the effects of oppression on valued life-outcomes have employed an intersectional approach to more fully represent a person’s identity and capture the multiple, and often compound, impacts of oppression. The study presented here replicated an intersectional approach to DIF analyses to examine whether findings from a previous study that focused on a single grade-level achievement test generalized to other subject areas and grade levels. Findings indicate that the use of an intersectional approach is more sensitive to detecting potential item bias and that this increased sensitivity holds across the subject areas and grade levels examined.
微分项目函数(DIF)分析通常用于检查由测试项目产生的潜在偏差。自引入以来,DIF分析一直侧重于与广泛压迫类别相关的潜在偏见,包括性别、种族分层、经济阶层和能力。最近,研究压迫对有价值的生活结果的影响的努力采用了一种交叉的方法,以更充分地代表一个人的身份,并捕捉到压迫的多重(通常是复合的)影响。本文提出的研究复制了DIF分析的交叉方法,以检验先前研究的结果是否适用于其他学科领域和年级水平。研究结果表明,使用交叉方法对发现潜在的项目偏见更敏感,而且这种增加的敏感性在被检查的学科领域和年级水平上都适用。
{"title":"An Intersectional Approach to DIF: Do Initial Findings Hold across Tests?","authors":"M. Russell, Olivia Szendey, Larry Kaplan","doi":"10.1080/10627197.2021.1965473","DOIUrl":"https://doi.org/10.1080/10627197.2021.1965473","url":null,"abstract":"ABSTRACT Differential Item Function (DIF) analysis is commonly employed to examine potential bias produced by a test item. Since its introduction DIF analyses have focused on potential bias related to broad categories of oppression, including gender, racial stratification, economic class, and ableness. More recently, efforts to examine the effects of oppression on valued life-outcomes have employed an intersectional approach to more fully represent a person’s identity and capture the multiple, and often compound, impacts of oppression. The study presented here replicated an intersectional approach to DIF analyses to examine whether findings from a previous study that focused on a single grade-level achievement test generalized to other subject areas and grade levels. Findings indicate that the use of an intersectional approach is more sensitive to detecting potential item bias and that this increased sensitivity holds across the subject areas and grade levels examined.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42202501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Beyond Agreement: Exploring Rater Effects in Large-Scale Mixed Format Assessments 超越协议:探索大规模混合格式评估中的评级效应
IF 1.5 Q2 Social Sciences Pub Date : 2021-08-17 DOI: 10.1080/10627197.2021.1962277
Stefanie A. Wind, Wenjing Guo
ABSTRACT Scoring procedures for the constructed-response (CR) items in large-scale mixed-format educational assessments often involve checks for rater agreement or rater reliability. Although these analyses are important, researchers have documented rater effects that persist despite rater training and that are not always detected in rater agreement and reliability analyses, such as severity/leniency, centrality/extremism, and biases. Left undetected, these effects pose threats to fairness. We illustrate how rater effects analyses can be incorporated into scoring procedures for large-scale mixed-format assessments. We used data from the National Assessment of Educational Progress (NAEP) to illustrate relatively simple analyses that can provide insight into patterns of rater judgment that may warrant additional attention. Our results suggested that the NAEP raters exhibited generally defensible psychometric properties, while also exhibiting some idiosyncrasies that could inform scoring procedures. Similar procedures could be used operationally to inform the interpretation and use of rater judgments in large-scale mixed-format assessments.
在大规模混合格式教育评估中,构建反应(CR)项目的评分程序通常包括对评分者一致性或可靠性的检查。尽管这些分析很重要,但研究人员已经证明,尽管进行了评级培训,但评级效应仍然存在,并且在评级一致性和可靠性分析中并不总是检测到,例如严厉/宽大、中心性/极端主义和偏见。如果不被发现,这些影响会对公平构成威胁。我们说明了如何将评分者效应分析纳入大规模混合格式评估的评分程序。我们使用来自国家教育进步评估(NAEP)的数据来说明相对简单的分析,这些分析可以提供对评分判断模式的洞察,这些模式可能需要额外的关注。我们的研究结果表明,NAEP评分者表现出一般可辩护的心理测量特性,同时也表现出一些可以为评分程序提供信息的特质。类似的程序可在业务上用于大规模混合格式评估中解释和使用较重的判断。
{"title":"Beyond Agreement: Exploring Rater Effects in Large-Scale Mixed Format Assessments","authors":"Stefanie A. Wind, Wenjing Guo","doi":"10.1080/10627197.2021.1962277","DOIUrl":"https://doi.org/10.1080/10627197.2021.1962277","url":null,"abstract":"ABSTRACT Scoring procedures for the constructed-response (CR) items in large-scale mixed-format educational assessments often involve checks for rater agreement or rater reliability. Although these analyses are important, researchers have documented rater effects that persist despite rater training and that are not always detected in rater agreement and reliability analyses, such as severity/leniency, centrality/extremism, and biases. Left undetected, these effects pose threats to fairness. We illustrate how rater effects analyses can be incorporated into scoring procedures for large-scale mixed-format assessments. We used data from the National Assessment of Educational Progress (NAEP) to illustrate relatively simple analyses that can provide insight into patterns of rater judgment that may warrant additional attention. Our results suggested that the NAEP raters exhibited generally defensible psychometric properties, while also exhibiting some idiosyncrasies that could inform scoring procedures. Similar procedures could be used operationally to inform the interpretation and use of rater judgments in large-scale mixed-format assessments.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48562788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Investigating the Use of Assessment Data by Primary School Teachers: Insights from a Large-scale Survey in Ireland 调查小学教师评估数据的使用:来自爱尔兰大规模调查的见解
IF 1.5 Q2 Social Sciences Pub Date : 2021-07-03 DOI: 10.1080/10627197.2021.1917358
Vasiliki Pitsia, Anastasios Karakolidis, P. Lehane
ABSTRACT Evidence suggests that the quality of teachers’ instructional practices can be improved when these are informed by relevant assessment data. Drawing on a sample of 1,300 primary school teachers in Ireland, this study examined the extent to which teachers use standardized test results for instructional purposes as well as the role of several factors in predicting this use. Specifically, the study analyzed data from a cross-sectional survey that gathered information about teachers’ use of, experiences with, and attitudes toward assessment data from standardized tests. After taking other teacher and school characteristics into consideration, the analysis revealed that teachers with more positive attitudes toward standardized tests and those who were often engaged in some form of professional development on standardized testing tended to use assessment data to inform their teaching more frequently. Based on the findings, policy and practice implications are discussed.
摘要证据表明,当教师的教学实践质量得到相关评估数据的支持时,这些质量可以得到提高。这项研究以爱尔兰1300名小学教师为样本,考察了教师将标准化考试结果用于教学目的的程度,以及几个因素在预测这种使用中的作用。具体而言,该研究分析了一项横断面调查的数据,该调查收集了教师对标准化考试评估数据的使用、经验和态度的信息。在考虑了其他教师和学校的特点后,分析显示,对标准化考试态度更积极的教师,以及那些经常在标准化考试中进行某种形式的专业发展的教师,倾向于更频繁地使用评估数据来告知他们的教学。根据调查结果,讨论了政策和实践意义。
{"title":"Investigating the Use of Assessment Data by Primary School Teachers: Insights from a Large-scale Survey in Ireland","authors":"Vasiliki Pitsia, Anastasios Karakolidis, P. Lehane","doi":"10.1080/10627197.2021.1917358","DOIUrl":"https://doi.org/10.1080/10627197.2021.1917358","url":null,"abstract":"ABSTRACT Evidence suggests that the quality of teachers’ instructional practices can be improved when these are informed by relevant assessment data. Drawing on a sample of 1,300 primary school teachers in Ireland, this study examined the extent to which teachers use standardized test results for instructional purposes as well as the role of several factors in predicting this use. Specifically, the study analyzed data from a cross-sectional survey that gathered information about teachers’ use of, experiences with, and attitudes toward assessment data from standardized tests. After taking other teacher and school characteristics into consideration, the analysis revealed that teachers with more positive attitudes toward standardized tests and those who were often engaged in some form of professional development on standardized testing tended to use assessment data to inform their teaching more frequently. Based on the findings, policy and practice implications are discussed.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/10627197.2021.1917358","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42751021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Using Full-information Item Analysis to Improve Item Quality 利用全信息项目分析提高项目质量
IF 1.5 Q2 Social Sciences Pub Date : 2021-07-03 DOI: 10.1080/10627197.2021.1946390
T. Haladyna, Michael C. Rodriguez
ABSTRACT Full-information item analysis provides item developers and reviewers comprehensive empirical evidence of item quality, including option response frequency, point-biserial index (PBI) for distractors, mean-scores of respondents selecting each option, and option trace lines. The multi-serial index (MSI) is introduced as a more informative item-total correlation, accounting for variable distractor performance. The overall item PBI is empirically compared to the MSI. For items from an operational mathematics and reading test, poorly performing distractors are systematically removed to recompute the MSI, indicating improvements in item quality. Case studies for specific items with different characteristics are described to illustrate a variety of outcomes, focused on improving item discrimination. Full-information item analyses are presented for each case study item, providing clear examples of interpretation and use of item analyses. A summary of recommendations for item analysts is provided.
全信息项目分析为项目的开发者和评论者提供了项目质量的综合经验证据,包括选项响应频率、干扰因素的点双列指数(PBI)、选择每个选项的被调查者的平均得分以及选项跟踪线。多序列指数(MSI)被引入作为一个更信息的项目-总相关,考虑到可变的干扰性能。总体项目PBI与MSI进行经验比较。对于运算数学和阅读测试中的项目,系统地移除表现不佳的干扰物以重新计算MSI,表明项目质量的改善。描述了具有不同特征的特定项目的案例研究,以说明各种结果,重点是改善项目歧视。为每个案例研究项目提供了完整的信息项目分析,提供了解释和使用项目分析的清晰示例。提供了对项目分析人员的建议摘要。
{"title":"Using Full-information Item Analysis to Improve Item Quality","authors":"T. Haladyna, Michael C. Rodriguez","doi":"10.1080/10627197.2021.1946390","DOIUrl":"https://doi.org/10.1080/10627197.2021.1946390","url":null,"abstract":"ABSTRACT Full-information item analysis provides item developers and reviewers comprehensive empirical evidence of item quality, including option response frequency, point-biserial index (PBI) for distractors, mean-scores of respondents selecting each option, and option trace lines. The multi-serial index (MSI) is introduced as a more informative item-total correlation, accounting for variable distractor performance. The overall item PBI is empirically compared to the MSI. For items from an operational mathematics and reading test, poorly performing distractors are systematically removed to recompute the MSI, indicating improvements in item quality. Case studies for specific items with different characteristics are described to illustrate a variety of outcomes, focused on improving item discrimination. Full-information item analyses are presented for each case study item, providing clear examples of interpretation and use of item analyses. A summary of recommendations for item analysts is provided.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/10627197.2021.1946390","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42811776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
The Impact of Disengaged Test Taking on a State’s Accountability Test Results 不参与考试对州问责制考试结果的影响
IF 1.5 Q2 Social Sciences Pub Date : 2021-07-03 DOI: 10.1080/10627197.2021.1956897
S. Wise, Sukkeun Im, Jay Lee
ABSTRACT This study investigated test-taking engagement on the Spring 2019 administration of a large-scale state summative assessment. Through the identification of rapid-guessing behavior – which is a validated indicator of disengagement – the percentage of Grade 8 test events with meaningful amounts of rapid guessing was 5.5% in mathematics, 6.7% in English Language Arts (ELA), and 3.5% in science. Disengagement rates on the state summative test were also found to vary materially across gender, ethnicity, Individualized Educational Plan (IEP) status, Limited English Proficient (LEP) status, free and reduced lunch (FRL) status, and disability status. However, school mean performance, proficiency rates, and relative ranking were only minimally affected by disengagement. Overall, results of this study indicate that disengagement has a material impact on individual state summative test scores, though its impact on score aggregations may be relatively minor.
摘要本研究调查了2019年春季大规模州总结性评估的参与情况。通过识别快速猜测行为(这是一个有效的脱离接触指标),8年级测试中有意义的快速猜测的百分比在数学中为5.5%,在英语语言艺术(ELA)中为6.7%,在科学中为3.5%。州总结性考试的脱离率也因性别、种族、个性化教育计划(IEP)状态、英语能力有限(LEP)状态,免费和减少午餐(FRL)状态以及残疾状态而存在重大差异。然而,脱离接触对学校平均成绩、熟练率和相对排名的影响微乎其微。总的来说,这项研究的结果表明,脱离对个人状态总结性考试成绩有实质性影响,尽管它对成绩汇总的影响可能相对较小。
{"title":"The Impact of Disengaged Test Taking on a State’s Accountability Test Results","authors":"S. Wise, Sukkeun Im, Jay Lee","doi":"10.1080/10627197.2021.1956897","DOIUrl":"https://doi.org/10.1080/10627197.2021.1956897","url":null,"abstract":"ABSTRACT This study investigated test-taking engagement on the Spring 2019 administration of a large-scale state summative assessment. Through the identification of rapid-guessing behavior – which is a validated indicator of disengagement – the percentage of Grade 8 test events with meaningful amounts of rapid guessing was 5.5% in mathematics, 6.7% in English Language Arts (ELA), and 3.5% in science. Disengagement rates on the state summative test were also found to vary materially across gender, ethnicity, Individualized Educational Plan (IEP) status, Limited English Proficient (LEP) status, free and reduced lunch (FRL) status, and disability status. However, school mean performance, proficiency rates, and relative ranking were only minimally affected by disengagement. Overall, results of this study indicate that disengagement has a material impact on individual state summative test scores, though its impact on score aggregations may be relatively minor.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/10627197.2021.1956897","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45119734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Assessing Quality of Teaching from Different Perspectives: Measurement Invariance across Teachers and Classes 从不同角度评估教学质量:教师和班级之间的测量不变性
IF 1.5 Q2 Social Sciences Pub Date : 2021-04-03 DOI: 10.1080/10627197.2020.1858785
G. Krammer, Barbara Pflanzl, Gerlinde Lenske, Johannes Mayr
ABSTRACT Comparing teachers’ self-assessment to classes’ assessment of quality of teaching can offer insights for educational research and be a valuable resource for teachers’ continuous professional development. However, the quality of teaching needs to be measured in the same way across perspectives for this comparison to be meaningful. We used data from 622 teachers self-assessing aspects of quality of teaching and of their classes (12229 students) assessing the same aspects. Perspectives were compared with measurement invariance analyses. Teachers and classes agreed on the average level of instructional clarity, and disagreed over teacher-student relationship and performance monitoring, suggesting that mean differences across perspectives may not be as consistent as the literature claims. Results showed a nonuniform measurement bias for only one item of instructional clarity, while measurement of the other aspects was directly comparable. We conclude the viability of comparing teachers’ and classes’ perspectives of aspects of quality of teaching.
将教师自我评价与课堂教学质量评价进行比较,可以为教育研究提供见解,是教师持续专业发展的宝贵资源。然而,为了使这种比较有意义,教学质量需要以同样的方式跨视角进行衡量。我们使用的数据来自622名教师对教学质量的自我评估,以及他们的班级(12229名学生)对相同方面的评估。比较透视与测量不变性分析。教师和班级在教学清晰度的平均水平上达成一致,但在师生关系和绩效监控方面存在分歧,这表明不同观点的平均差异可能不像文献所声称的那样一致。结果显示,只有一个项目的教学清晰度不均匀的测量偏差,而测量的其他方面是直接可比。我们总结了比较教师和学生在教学质量方面的观点的可行性。
{"title":"Assessing Quality of Teaching from Different Perspectives: Measurement Invariance across Teachers and Classes","authors":"G. Krammer, Barbara Pflanzl, Gerlinde Lenske, Johannes Mayr","doi":"10.1080/10627197.2020.1858785","DOIUrl":"https://doi.org/10.1080/10627197.2020.1858785","url":null,"abstract":"ABSTRACT Comparing teachers’ self-assessment to classes’ assessment of quality of teaching can offer insights for educational research and be a valuable resource for teachers’ continuous professional development. However, the quality of teaching needs to be measured in the same way across perspectives for this comparison to be meaningful. We used data from 622 teachers self-assessing aspects of quality of teaching and of their classes (12229 students) assessing the same aspects. Perspectives were compared with measurement invariance analyses. Teachers and classes agreed on the average level of instructional clarity, and disagreed over teacher-student relationship and performance monitoring, suggesting that mean differences across perspectives may not be as consistent as the literature claims. Results showed a nonuniform measurement bias for only one item of instructional clarity, while measurement of the other aspects was directly comparable. We conclude the viability of comparing teachers’ and classes’ perspectives of aspects of quality of teaching.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-04-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/10627197.2020.1858785","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42517403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Predicting Retention in Higher Education from high-stakes Exams or School GPA 从高风险考试或学校平均成绩预测高等教育的保留率
IF 1.5 Q2 Social Sciences Pub Date : 2021-02-06 DOI: 10.1080/10627197.2022.2130748
M. Meeter, M. V. van Brederode
ABSTRACT The transition from secondary to tertiary education varies from country to country. In many countries, secondary school is concluded with high-stakes, national exams, or high-stakes entry tests are used for admissions to tertiary education. In other countries, secondary-school grade point average (GPA) is the determining factor. In the Netherlands, both play a role. With administrative data of close to 180,000 students, we investigated whether national exam scores or secondary school GPA was a better predictor of tertiary first-year retention. For both university education and higher professional education, secondary school GPA was the better prediction of retention, to the extent that national exams did not explain any additional variance. Moreover, for students who failed their exam, being held back by the secondary school for an additional year and entering tertiary education one year later, GPA in the year of failure remained as predictive as for students who had passed their exams and started tertiary education immediately. National exam scores, on the other hand, had no predictive value at all for these students. It is concluded that secondary school GPA measures aspects of student performance that is not included in high-stakes national exams, but that are predictive of subsequent success in tertiary education.
从中等教育到高等教育的转变因国家而异。在许多国家,中学是通过高风险的国家考试结束的,或者高风险的入学考试用于进入高等教育。在其他国家,中学平均绩点(GPA)是决定因素。在荷兰,两者都发挥了作用。利用近18万名学生的行政数据,我们调查了国家考试成绩或中学GPA是否能更好地预测大学一年级的留校率。无论是大学教育还是高等职业教育,中学GPA都能更好地预测留校率,在某种程度上,国家考试不能解释任何额外的差异。此外,对于那些考试不及格的学生,他们被中学多留了一年,一年后进入高等教育,他们在考试不及格那年的GPA仍然与通过考试并立即开始高等教育的学生一样具有预测性。另一方面,国家考试成绩对这些学生没有任何预测价值。结论是,中学GPA衡量的是学生表现的一些方面,这些方面不包括在高风险的国家考试中,但却预示着随后在高等教育中的成功。
{"title":"Predicting Retention in Higher Education from high-stakes Exams or School GPA","authors":"M. Meeter, M. V. van Brederode","doi":"10.1080/10627197.2022.2130748","DOIUrl":"https://doi.org/10.1080/10627197.2022.2130748","url":null,"abstract":"ABSTRACT The transition from secondary to tertiary education varies from country to country. In many countries, secondary school is concluded with high-stakes, national exams, or high-stakes entry tests are used for admissions to tertiary education. In other countries, secondary-school grade point average (GPA) is the determining factor. In the Netherlands, both play a role. With administrative data of close to 180,000 students, we investigated whether national exam scores or secondary school GPA was a better predictor of tertiary first-year retention. For both university education and higher professional education, secondary school GPA was the better prediction of retention, to the extent that national exams did not explain any additional variance. Moreover, for students who failed their exam, being held back by the secondary school for an additional year and entering tertiary education one year later, GPA in the year of failure remained as predictive as for students who had passed their exams and started tertiary education immediately. National exam scores, on the other hand, had no predictive value at all for these students. It is concluded that secondary school GPA measures aspects of student performance that is not included in high-stakes national exams, but that are predictive of subsequent success in tertiary education.","PeriodicalId":46209,"journal":{"name":"Educational Assessment","volume":null,"pages":null},"PeriodicalIF":1.5,"publicationDate":"2021-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42962496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Educational Assessment
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1