首页 > 最新文献

Journal of Educational and Behavioral Statistics最新文献

英文 中文
Monitoring Item Performance With CUSUM Statistics in Continuous Testing 在连续测试中使用CUSUM统计监测项目性能
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-03-08 DOI: 10.3102/1076998621994563
Yi-Hsuan Lee, C. Lewis
In many educational assessments, items are reused in different administrations throughout the life of the assessments. Ideally, a reused item should perform relatively similarly over time. In reality, an item may become easier with exposure, especially when item preknowledge has occurred. This article presents a novel cumulative sum procedure for detecting item preknowledge in continuous testing where data for each reused item may be obtained from small and varying sample sizes across administrations. Its performance is evaluated with simulations and analytical work. The approach is effective in detecting item preknowledge quickly with group size at least 10 and is easy to implement with varying item parameters. In addition, it is robust to the ability estimation error introduced in the simulations.
在许多教育评估中,项目在评估的整个生命周期内被不同的管理部门重复使用。理想情况下,随着时间的推移,重复使用的项目应该表现得相对相似。事实上,一个项目可能会随着曝光而变得更容易,尤其是当项目发生预先知情时。本文提出了一种新的累积和程序,用于在连续测试中检测项目先验知识,其中每个重复使用项目的数据可以从不同管理部门的小样本量和不同样本量中获得。通过模拟和分析工作对其性能进行了评估。该方法在组大小至少为10的情况下快速检测项目先验知识是有效的,并且在不同项目参数的情况下易于实现。此外,它对仿真中引入的能力估计误差具有鲁棒性。
{"title":"Monitoring Item Performance With CUSUM Statistics in Continuous Testing","authors":"Yi-Hsuan Lee, C. Lewis","doi":"10.3102/1076998621994563","DOIUrl":"https://doi.org/10.3102/1076998621994563","url":null,"abstract":"In many educational assessments, items are reused in different administrations throughout the life of the assessments. Ideally, a reused item should perform relatively similarly over time. In reality, an item may become easier with exposure, especially when item preknowledge has occurred. This article presents a novel cumulative sum procedure for detecting item preknowledge in continuous testing where data for each reused item may be obtained from small and varying sample sizes across administrations. Its performance is evaluated with simulations and analytical work. The approach is effective in detecting item preknowledge quickly with group size at least 10 and is easy to implement with varying item parameters. In addition, it is robust to the ability estimation error introduced in the simulations.","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"46 1","pages":"611 - 648"},"PeriodicalIF":2.4,"publicationDate":"2021-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48504334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Jenss–Bayley Latent Change Score Model With Individual Ratio of the Growth Acceleration in the Framework of Individual Measurement Occasions 个体测量情景框架下具有个体增长加速率的Jens–Bayley潜在变化得分模型
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-02-27 DOI: 10.3102/10769986221099919
Jin Liu
Longitudinal data analysis has been widely employed to examine between-individual differences in within-individual changes. One challenge of such analyses is that the rate-of-change is only available indirectly when change patterns are nonlinear with respect to time. Latent change score models (LCSMs), which can be employed to investigate the change in rate-of-change at the individual level, have been developed to address this challenge. We extend an existing LCSM with the Jenss–Bayley growth curve and propose a novel expression for change scores that allows for (1) unequally spaced study waves and (2) individual measurement occasions around each wave. We also extend the existing model to estimate the individual ratio of the growth acceleration (that largely determines the trajectory shape and is viewed as the most important parameter in the Jenss–Bayley model). We present the proposed model by a simulation study and a real-world data analysis. Our simulation study demonstrates that the proposed model can estimate the parameters unbiasedly and precisely and exhibit target confidence interval coverage. The simulation study also shows that the proposed model with the novel expression for the change scores outperforms the existing model. An empirical example using longitudinal reading scores shows that the model can estimate the individual ratio of the growth acceleration and generate individual rate-of-change in practice. We also provide the corresponding code for the proposed model.
纵向数据分析已被广泛用于检验个体内部变化中的个体间差异。这种分析的一个挑战是,只有当变化模式相对于时间是非线性的时候,变化率才能间接得到。为了应对这一挑战,人们开发了潜在变化评分模型(lcsm),该模型可用于研究个人水平上的变化率变化。我们用Jenss-Bayley生长曲线扩展了现有的LCSM,并提出了一种新的变化分数表达式,该表达式允许(1)不均匀间隔的学习波和(2)每个波周围的单独测量场合。我们还扩展了现有模型来估计生长加速的个体比率(这在很大程度上决定了轨迹形状,被视为Jenss-Bayley模型中最重要的参数)。我们通过仿真研究和实际数据分析提出了该模型。仿真研究表明,该模型能够准确、无偏地估计参数,并具有目标置信区间覆盖率。仿真研究还表明,采用新的变化分数表达式的模型优于现有的模型。通过纵向阅读分数的实证分析表明,该模型在实际应用中能够较好地估计个体的增长加速比,并生成个体的变化速率。我们还为提议的模型提供了相应的代码。
{"title":"Jenss–Bayley Latent Change Score Model With Individual Ratio of the Growth Acceleration in the Framework of Individual Measurement Occasions","authors":"Jin Liu","doi":"10.3102/10769986221099919","DOIUrl":"https://doi.org/10.3102/10769986221099919","url":null,"abstract":"Longitudinal data analysis has been widely employed to examine between-individual differences in within-individual changes. One challenge of such analyses is that the rate-of-change is only available indirectly when change patterns are nonlinear with respect to time. Latent change score models (LCSMs), which can be employed to investigate the change in rate-of-change at the individual level, have been developed to address this challenge. We extend an existing LCSM with the Jenss–Bayley growth curve and propose a novel expression for change scores that allows for (1) unequally spaced study waves and (2) individual measurement occasions around each wave. We also extend the existing model to estimate the individual ratio of the growth acceleration (that largely determines the trajectory shape and is viewed as the most important parameter in the Jenss–Bayley model). We present the proposed model by a simulation study and a real-world data analysis. Our simulation study demonstrates that the proposed model can estimate the parameters unbiasedly and precisely and exhibit target confidence interval coverage. The simulation study also shows that the proposed model with the novel expression for the change scores outperforms the existing model. An empirical example using longitudinal reading scores shows that the model can estimate the individual ratio of the growth acceleration and generate individual rate-of-change in practice. We also provide the corresponding code for the proposed model.","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"47 1","pages":"507 - 543"},"PeriodicalIF":2.4,"publicationDate":"2021-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47673583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Estimating Difference-Score Reliability in Pretest–Posttest Settings 评估测试前-测试后设置中的差异得分可靠性
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-02-15 DOI: 10.3102/1076998620986948
Zhengguo Gu, W. Emons, K. Sijtsma
Clinical, medical, and health psychologists use difference scores obtained from pretest–posttest designs employing the same test to assess intraindividual change possibly caused by an intervention addressing, for example, anxiety, depression, eating disorder, or addiction. Reliability of difference scores is important for interpreting observed change. This article compares the well-documented traditional method and the unfamiliar, rarely used item-level method for estimating difference-score reliability. We simulated data under various conditions that are typical of change assessment in pretest–posttest designs. The item-level method had smaller bias and greater precision than the traditional method and may be recommended for practical use.
临床、医学和健康心理学家使用从采用相同测试的前测-后测设计中获得的差异分数来评估可能由针对焦虑、抑郁、饮食障碍或成瘾的干预措施引起的个体内变化。差异得分的可靠性对于解释观察到的变化很重要。本文比较了文献丰富的传统方法和不熟悉、很少使用的项目级方法来估计差异得分的可靠性。我们模拟了各种条件下的数据,这些条件是前测-后测设计中变化评估的典型情况。与传统方法相比,项目级方法具有更小的偏差和更高的精度,可以推荐用于实际应用。
{"title":"Estimating Difference-Score Reliability in Pretest–Posttest Settings","authors":"Zhengguo Gu, W. Emons, K. Sijtsma","doi":"10.3102/1076998620986948","DOIUrl":"https://doi.org/10.3102/1076998620986948","url":null,"abstract":"Clinical, medical, and health psychologists use difference scores obtained from pretest–posttest designs employing the same test to assess intraindividual change possibly caused by an intervention addressing, for example, anxiety, depression, eating disorder, or addiction. Reliability of difference scores is important for interpreting observed change. This article compares the well-documented traditional method and the unfamiliar, rarely used item-level method for estimating difference-score reliability. We simulated data under various conditions that are typical of change assessment in pretest–posttest designs. The item-level method had smaller bias and greater precision than the traditional method and may be recommended for practical use.","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"46 1","pages":"592 - 610"},"PeriodicalIF":2.4,"publicationDate":"2021-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46183979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Statistical Power for Estimating Treatment Effects Using Difference-in-Differences and Comparative Interrupted Time Series Estimators With Variation in Treatment Timing 利用差中之差和具有治疗时间变化的比较中断时间序列估计器估计治疗效果的统计能力
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-02-12 DOI: 10.3102/10769986211070625
Peter Z. Schochet
This article develops new closed-form variance expressions for power analyses for commonly used difference-in-differences (DID) and comparative interrupted time series (CITS) panel data estimators. The main contribution is to incorporate variation in treatment timing into the analysis. The power formulas also account for other key design features that arise in practice: autocorrelated errors, unequal measurement intervals, and clustering due to the unit of treatment assignment. We consider power formulas for both cross-sectional and longitudinal models and allow for covariates. An illustrative power analysis provides guidance on appropriate sample sizes. The key finding is that accounting for treatment timing increases required sample sizes. Further, DID estimators have considerably more power than standard CITS and ITS estimators. An available Shiny R dashboard performs the sample size calculations for the considered estimators.
本文为常用的差分(DID)和比较中断时间序列(CITS)面板数据估计量开发了新的幂分析闭式方差表达式。主要贡献是将治疗时间的变化纳入分析。幂公式还考虑了实践中出现的其他关键设计特征:自相关误差、不相等的测量间隔以及由于治疗分配单位而导致的聚类。我们考虑横截面和纵向模型的幂公式,并考虑协变。说明性的功率分析提供了关于适当样本大小的指导。关键发现是,考虑到治疗时间会增加所需的样本量。此外,DID估计器比标准CITS和ITS估计器具有更大的功率。一个可用的Shiny R仪表板为所考虑的估计量执行样本量计算。
{"title":"Statistical Power for Estimating Treatment Effects Using Difference-in-Differences and Comparative Interrupted Time Series Estimators With Variation in Treatment Timing","authors":"Peter Z. Schochet","doi":"10.3102/10769986211070625","DOIUrl":"https://doi.org/10.3102/10769986211070625","url":null,"abstract":"This article develops new closed-form variance expressions for power analyses for commonly used difference-in-differences (DID) and comparative interrupted time series (CITS) panel data estimators. The main contribution is to incorporate variation in treatment timing into the analysis. The power formulas also account for other key design features that arise in practice: autocorrelated errors, unequal measurement intervals, and clustering due to the unit of treatment assignment. We consider power formulas for both cross-sectional and longitudinal models and allow for covariates. An illustrative power analysis provides guidance on appropriate sample sizes. The key finding is that accounting for treatment timing increases required sample sizes. Further, DID estimators have considerably more power than standard CITS and ITS estimators. An available Shiny R dashboard performs the sample size calculations for the considered estimators.","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"47 1","pages":"367 - 405"},"PeriodicalIF":2.4,"publicationDate":"2021-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41400358","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Computational Strategies and Estimation Performance With Bayesian Semiparametric Item Response Theory Models 贝叶斯半参数项目反应理论模型的计算策略和估计性能
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-01-27 DOI: 10.3102/10769986221136105
S. Paganin, C. Paciorek, Claudia Wehrhahn, Abel Rodríguez, S. Rabe-Hesketh, P. de Valpine
Item response theory (IRT) models typically rely on a normality assumption for subject-specific latent traits, which is often unrealistic in practice. Semiparametric extensions based on Dirichlet process mixtures (DPMs) offer a more flexible representation of the unknown distribution of the latent trait. However, the use of such models in the IRT literature has been extremely limited, in good part because of the lack of comprehensive studies and accessible software tools. This article provides guidance for practitioners on semiparametric IRT models and their implementation. In particular, we rely on NIMBLE, a flexible software system for hierarchical models that enables the use of DPMs. We highlight efficient sampling strategies for model estimation and compare inferential results under parametric and semiparametric models.
项目反应理论(IRT)模型通常依赖于对特定主题潜在特征的正态性假设,这在实践中往往是不现实的。基于狄利克雷过程混合物(DPM)的半参数扩展为潜在特征的未知分布提供了更灵活的表示。然而,IRT文献中对此类模型的使用极为有限,这在很大程度上是因为缺乏全面的研究和可访问的软件工具。本文为从业者提供了关于半参数IRT模型及其实现的指导。特别是,我们依赖NIMBLE,这是一个用于分层模型的灵活软件系统,可以使用DPM。我们强调了模型估计的有效采样策略,并比较了参数和半参数模型下的推断结果。
{"title":"Computational Strategies and Estimation Performance With Bayesian Semiparametric Item Response Theory Models","authors":"S. Paganin, C. Paciorek, Claudia Wehrhahn, Abel Rodríguez, S. Rabe-Hesketh, P. de Valpine","doi":"10.3102/10769986221136105","DOIUrl":"https://doi.org/10.3102/10769986221136105","url":null,"abstract":"Item response theory (IRT) models typically rely on a normality assumption for subject-specific latent traits, which is often unrealistic in practice. Semiparametric extensions based on Dirichlet process mixtures (DPMs) offer a more flexible representation of the unknown distribution of the latent trait. However, the use of such models in the IRT literature has been extremely limited, in good part because of the lack of comprehensive studies and accessible software tools. This article provides guidance for practitioners on semiparametric IRT models and their implementation. In particular, we rely on NIMBLE, a flexible software system for hierarchical models that enables the use of DPMs. We highlight efficient sampling strategies for model estimation and compare inferential results under parametric and semiparametric models.","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"48 1","pages":"147 - 188"},"PeriodicalIF":2.4,"publicationDate":"2021-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43423727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Regression Discontinuity Designs With an Ordinal Running Variable: Evaluating the Effects of Extended Time Accommodations for English-Language Learners 具有有序运行变量的回归不连续设计:评估英语学习者延长住宿时间的影响
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-01-20 DOI: 10.3102/10769986221090275
Youmi Suk, Peter M Steiner, Jee-Seon Kim, Hyunseung Kang
Regression discontinuity (RD) designs are commonly used for program evaluation with continuous treatment assignment variables. But in practice, treatment assignment is frequently based on ordinal variables. In this study, we propose an RD design with an ordinal running variable to assess the effects of extended time accommodations (ETA) for English-language learners (ELLs). ETA eligibility is determined by ordinal ELL English-proficiency categories of National Assessment of Educational Progress data. We discuss the identification and estimation of the average treatment effect (ATE), intent-to-treat effect, and the local ATE at the cutoff. We also propose a series of sensitivity analyses to probe the effect estimates’ robustness to the choices of scaling functions and cutoff scores and remaining confounding.
回归不连续(RD)设计通常用于具有连续处理分配变量的方案评估。但在实践中,治疗分配往往是基于有序变量。在这项研究中,我们提出了一个带有有序运行变量的RD设计来评估延长时间住宿(ETA)对英语学习者(ELLs)的影响。ETA资格由国家教育进步评估数据的有序英语水平类别决定。我们讨论了识别和估计平均处理效果(ATE),意向治疗效果,和局部ATE在截止点。我们还提出了一系列敏感性分析,以探讨效果估计对尺度函数和截止分数的选择和剩余混淆的稳健性。
{"title":"Regression Discontinuity Designs With an Ordinal Running Variable: Evaluating the Effects of Extended Time Accommodations for English-Language Learners","authors":"Youmi Suk, Peter M Steiner, Jee-Seon Kim, Hyunseung Kang","doi":"10.3102/10769986221090275","DOIUrl":"https://doi.org/10.3102/10769986221090275","url":null,"abstract":"Regression discontinuity (RD) designs are commonly used for program evaluation with continuous treatment assignment variables. But in practice, treatment assignment is frequently based on ordinal variables. In this study, we propose an RD design with an ordinal running variable to assess the effects of extended time accommodations (ETA) for English-language learners (ELLs). ETA eligibility is determined by ordinal ELL English-proficiency categories of National Assessment of Educational Progress data. We discuss the identification and estimation of the average treatment effect (ATE), intent-to-treat effect, and the local ATE at the cutoff. We also propose a series of sensitivity analyses to probe the effect estimates’ robustness to the choices of scaling functions and cutoff scores and remaining confounding.","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"47 1","pages":"459 - 484"},"PeriodicalIF":2.4,"publicationDate":"2021-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44101445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cross-Classified Random Effects Modeling for Moderated Item Calibration 适度项目校准的交叉分类随机效应模型
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2021-01-12 DOI: 10.3102/1076998620983908
Seungwon Chung, Li Cai
In the research reported here, we propose a new method for scale alignment and test scoring in the context of supporting students with disabilities. In educational assessment, students from these special populations take modified tests because of a demonstrated disability that requires more assistance than standard testing accommodation. Updated federal education legislation and guidance require that these students be assessed and included in state education accountability systems, and their achievement reported with respect to the same rigorous content and achievement standards that the state adopted. Routine item calibration and linking methods are not feasible because the size of these special populations tends to be small. We develop a unified cross-classified random effects model that utilizes item response data from the general population as well as judge-provided data from subject matter experts in order to obtain revised item parameter estimates for use in scoring modified tests. We extend the Metropolis–Hastings Robbins–Monro algorithm to estimate the parameters of this model. The proposed method is applied to Braille test forms in a large operational multistate English language proficiency assessment program. Our work not only allows a broader range of modifications that is routinely considered in large-scale educational assessments but also directly incorporates the input from subject matter experts who work directly with the students needing support. Their structured and informed feedback deserves more attention from the psychometric community.
在本研究中,我们提出了一种在支持残疾学生的背景下进行量表校准和测试评分的新方法。在教育评估中,来自这些特殊群体的学生参加修改后的考试,因为他们证明有残疾,需要比标准考试便利更多的帮助。最新的联邦教育立法和指导要求对这些学生进行评估,并将其纳入州教育问责制,他们的成绩报告与州采用的严格内容和成绩标准相同。常规的项目校准和连接方法是不可行的,因为这些特殊人群的规模往往很小。我们开发了一个统一的交叉分类随机效应模型,该模型利用了来自一般人群的项目反应数据以及来自主题专家的判断提供的数据,以便获得用于评分修改测试的修订项目参数估计。我们扩展了Metropolis-Hastings - Robbins-Monro算法来估计该模型的参数。该方法已应用于一个大型多州英语语言能力评估项目的盲文测试表格。我们的工作不仅允许在大规模教育评估中常规考虑的更广泛的修改,而且还直接纳入了直接与需要支持的学生一起工作的主题专家的输入。他们的结构化和知情反馈值得心理测量界更多的关注。
{"title":"Cross-Classified Random Effects Modeling for Moderated Item Calibration","authors":"Seungwon Chung, Li Cai","doi":"10.3102/1076998620983908","DOIUrl":"https://doi.org/10.3102/1076998620983908","url":null,"abstract":"In the research reported here, we propose a new method for scale alignment and test scoring in the context of supporting students with disabilities. In educational assessment, students from these special populations take modified tests because of a demonstrated disability that requires more assistance than standard testing accommodation. Updated federal education legislation and guidance require that these students be assessed and included in state education accountability systems, and their achievement reported with respect to the same rigorous content and achievement standards that the state adopted. Routine item calibration and linking methods are not feasible because the size of these special populations tends to be small. We develop a unified cross-classified random effects model that utilizes item response data from the general population as well as judge-provided data from subject matter experts in order to obtain revised item parameter estimates for use in scoring modified tests. We extend the Metropolis–Hastings Robbins–Monro algorithm to estimate the parameters of this model. The proposed method is applied to Braille test forms in a large operational multistate English language proficiency assessment program. Our work not only allows a broader range of modifications that is routinely considered in large-scale educational assessments but also directly incorporates the input from subject matter experts who work directly with the students needing support. Their structured and informed feedback deserves more attention from the psychometric community.","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"46 1","pages":"651 - 681"},"PeriodicalIF":2.4,"publicationDate":"2021-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45633091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Practical Guide for Analyzing Large-Scale Assessment Data Using Mplus: A Case Demonstration Using the Program for International Assessment of Adult Competencies Data 使用Mplus分析大规模评估数据的实用指南:使用成人能力数据国际评估计划的案例演示
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2020-12-16 DOI: 10.3102/1076998620978554
T. Yamashita, Thomas J. Smith, P. Cummins
In order to promote the use of increasingly available large-scale assessment data in education and expand the scope of analytic capabilities among applied researchers, this study provides step-by-step guidance, and practical examples of syntax and data analysis using Mplus. Concise overview and key unique aspects of large-scale assessment data from the 2012/2014 Program for International Assessment of Adult Competencies (PIAAC) are described. Using commonly-used statistical software including SAS and R, a simple macro program and syntax are developed to streamline the data preparation process. Then, two examples of structural equation models are demonstrated using Mplus. The suggested data preparation and analytic approaches can be immediately applicable to existing large-scale assessment data.
为了促进在教育中使用越来越可用的大规模评估数据,并扩大应用研究人员的分析能力范围,本研究提供了使用Mplus进行语法和数据分析的分步指导和实际示例。简要概述了2012/2014年国际成人能力评估计划(PIAC)的大规模评估数据的主要独特方面。使用包括SAS和R在内的常用统计软件,开发了一个简单的宏程序和语法,以简化数据准备过程。然后,使用Mplus演示了结构方程模型的两个例子。建议的数据准备和分析方法可以立即适用于现有的大规模评估数据。
{"title":"A Practical Guide for Analyzing Large-Scale Assessment Data Using Mplus: A Case Demonstration Using the Program for International Assessment of Adult Competencies Data","authors":"T. Yamashita, Thomas J. Smith, P. Cummins","doi":"10.3102/1076998620978554","DOIUrl":"https://doi.org/10.3102/1076998620978554","url":null,"abstract":"In order to promote the use of increasingly available large-scale assessment data in education and expand the scope of analytic capabilities among applied researchers, this study provides step-by-step guidance, and practical examples of syntax and data analysis using Mplus. Concise overview and key unique aspects of large-scale assessment data from the 2012/2014 Program for International Assessment of Adult Competencies (PIAAC) are described. Using commonly-used statistical software including SAS and R, a simple macro program and syntax are developed to streamline the data preparation process. Then, two examples of structural equation models are demonstrated using Mplus. The suggested data preparation and analytic approaches can be immediately applicable to existing large-scale assessment data.","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"46 1","pages":"501 - 518"},"PeriodicalIF":2.4,"publicationDate":"2020-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.3102/1076998620978554","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43159243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Review of Handbook of Item Response Theory: Vol. 1 评《项目反应理论手册》第一卷
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2020-12-16 DOI: 10.3102/1076998620978551
Peter F. Halpin
The Handbook of Item Response Theory is an extensive three-volume collection with contributions from leading researchers in the field. This review focuses on Volume 1 (Models). Aside from the Introduction, each of the 33 chapters provides a self-contained presentation of an item response theory (IRT) modeling framework. The chapters share a common notation as well as a uniform organization (Introduction, Model Presentation, Parameter Estimation, Goodness of Fit, an Empirical Example, and a Discussion). Many chapters are leador singleauthored by original developers of the research, and in all cases, the lead authors are highly regarded as experts in the field. The Volume is organized into eight sections, each containing between two and seven chapters focused on types of data—dichotomous responses, polytomous responses, response times—or on types of models—multidimensional, nonparametric, nonmonotone, hierarchical and multilevel as well as generalized modeling approaches that include but are not limited to IRT applications. The coverage of models for polytomous data is especially strong, with seven chapters devoted to this topic. In other areas, the coverage is already appearing somewhat thin in light of recent research trends. For example, a large amount of work has been devoted to the analysis of response times since the publication of the Volume. The three chapters in the Volume provide the foundations of this more recent research, focusing the early work of Rasch, approaches based on cognitive models of decision making, and models for lognormal response times; the latter is extended to the joint modeling of responses and response times in a separate chapter. Generalized modeling approaches is another area that, in retrospect, could have received more thorough coverage of topics such as Bayesian IRT, psychometric applications of networks and graphs, or approaches based on machine learning. There is only one chapter addressing models with categorical latent variables. Despite the inevitable nit-picking about specific omissions, the Handbook certainly provides a thorough characterization of the breadth of active research on statistical models used in the IRT literature. Journal of Educational and Behavioral Statistics 2021, Vol. 46, No. 4, pp. 519–522 DOI: 10.3102/1076998620978551 Article reuse guidelines: sagepub.com/journals-permissions © 2020 AERA. http://jebs.aera.net
该手册的项目反应理论是一个广泛的三卷收集与贡献的主要研究人员在该领域。本综述的重点是第1卷(模型)。除了引言,33章中的每一章都提供了一个项目反应理论(IRT)建模框架的独立展示。章节共享一个共同的符号,以及一个统一的组织(介绍,模型表示,参数估计,拟合优度,一个经验的例子,并讨论)。许多章节都是由该研究的原始开发人员领导或单独撰写的,并且在所有情况下,主要作者都被视为该领域的专家。本卷分为八个部分,每个部分包含两到七章,重点介绍数据类型-二分类响应,多分类响应,响应时间-或模型类型-多维,非参数,非单调,分层和多层以及广义建模方法,包括但不限于IRT应用。多同构数据模型的覆盖面特别强,有七章专门讨论这个主题。在其他领域,鉴于最近的研究趋势,覆盖范围已经显得有些单薄。例如,自本卷出版以来,已经进行了大量的工作来分析响应时间。本卷的三章提供了这一最新研究的基础,重点是Rasch的早期工作,基于决策认知模型的方法,以及对数正态反应时间模型;后者将在单独的一章中扩展到响应和响应时间的联合建模。广义建模方法是另一个领域,回想起来,可以得到更全面的主题覆盖,如贝叶斯IRT,网络和图的心理测量应用,或基于机器学习的方法。只有一章讨论具有分类潜在变量的模型。尽管不可避免地会对具体的遗漏进行挑剔,但该手册确实提供了对IRT文献中使用的统计模型的活跃研究广度的全面描述。教育与行为统计杂志2021,Vol. 46, No. 4, pp. 519-522 DOI: 10.3102/1076998620978551文章重用指南:sagepub.com/journals-permissions©2020 AERA。http://jebs.aera.net
{"title":"A Review of Handbook of Item Response Theory: Vol. 1","authors":"Peter F. Halpin","doi":"10.3102/1076998620978551","DOIUrl":"https://doi.org/10.3102/1076998620978551","url":null,"abstract":"The Handbook of Item Response Theory is an extensive three-volume collection with contributions from leading researchers in the field. This review focuses on Volume 1 (Models). Aside from the Introduction, each of the 33 chapters provides a self-contained presentation of an item response theory (IRT) modeling framework. The chapters share a common notation as well as a uniform organization (Introduction, Model Presentation, Parameter Estimation, Goodness of Fit, an Empirical Example, and a Discussion). Many chapters are leador singleauthored by original developers of the research, and in all cases, the lead authors are highly regarded as experts in the field. The Volume is organized into eight sections, each containing between two and seven chapters focused on types of data—dichotomous responses, polytomous responses, response times—or on types of models—multidimensional, nonparametric, nonmonotone, hierarchical and multilevel as well as generalized modeling approaches that include but are not limited to IRT applications. The coverage of models for polytomous data is especially strong, with seven chapters devoted to this topic. In other areas, the coverage is already appearing somewhat thin in light of recent research trends. For example, a large amount of work has been devoted to the analysis of response times since the publication of the Volume. The three chapters in the Volume provide the foundations of this more recent research, focusing the early work of Rasch, approaches based on cognitive models of decision making, and models for lognormal response times; the latter is extended to the joint modeling of responses and response times in a separate chapter. Generalized modeling approaches is another area that, in retrospect, could have received more thorough coverage of topics such as Bayesian IRT, psychometric applications of networks and graphs, or approaches based on machine learning. There is only one chapter addressing models with categorical latent variables. Despite the inevitable nit-picking about specific omissions, the Handbook certainly provides a thorough characterization of the breadth of active research on statistical models used in the IRT literature. Journal of Educational and Behavioral Statistics 2021, Vol. 46, No. 4, pp. 519–522 DOI: 10.3102/1076998620978551 Article reuse guidelines: sagepub.com/journals-permissions © 2020 AERA. http://jebs.aera.net","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"46 1","pages":"519 - 522"},"PeriodicalIF":2.4,"publicationDate":"2020-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46999108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Acknowledgments 致谢
IF 2.4 3区 心理学 Q2 EDUCATION & EDUCATIONAL RESEARCH Pub Date : 2020-12-01 DOI: 10.3102/1076998620958383
C. Aßmann, Yinyin Chen, Bryan Keller, Jee-Seon Kim, K. Kim
Christian Aßmann, University of Bamberg Guillaume Basse, Stanford University Michela Battauz, University of Udine Eli Ben-Michael, University of CaliforniaBerkeley Howard Bloom, Manpower Demonstration Research Corporation (MDRC) Ulf Bockenholt, Northwestern University Daniel Bolt, University of Wisconsin Eric Bradlow, Wharton School of the University of Pennsylvania Henry Braun, Boston College Katherine Castellano, Educational Testing Service Wendy Chan, University of Pennsylvania Yinghan Chen, University of Nevada, Reno Yinyin Chen, University of Illinois at Urbana-Champaign Yunxiao Chen, London School of Economics and Political Science Edison Choe, Graduate Management Admission Council Steven Culpepper, University of Illinois at Urbana-Champaign Paul De Boeck, The Ohio State University Jimmy de la Torre, The University of Hong Kong Dries Debeer, University of Leuven (KU Leuven) Peng Ding, University of CaliforniaBerkeley Nianbo Dong, University of North Carolina at Chapel Hill Fritz Drasgow, University of Illinois at Urbana-Champaign Keelan Evanini, Educational Testing Service Jean-Paul Fox, University of Twente Mark Fredrickson, University of Michigan Jorge Gonzaléz, Pontificia Universidad Catolica de Chile Simon Grund, Leibniz-Institut fur die Padagogik der Naturwissenschaften und Mathematik an der Universitat Kiel Hongwen Guo, Educational Testing Service Gregory Hancock, University of MarylandCollege Park Ben Hansen, University of Michigan Jeffrey Harring, University of MarylandCollege Park Johannes Hartig, German Institute for International Educational Research (DIPF) Michael Harwell, University of Minnesota Timothy Hayes, Florida International University Yong He, ACT, Inc. Carolyn Hill, Manpower Demonstration Research Corporation (MDRC) Minjeong Jeon, University of CaliforniaLos Angeles Paul Jewsbury, Educational Testing Service Booil Jo, Stanford University Harry Joe, University of British Columbia Matthew Johnson, Educational Testing Service George Karabatsos, University of IllinoisChicago Luke Keele, University of Pennsylvania Augustin Kelava, Eberhard Karls University Tuebingen Journal of Educational and Behavioral Statistics 2020, Vol. 45, No. 6, pp. 771–773 DOI: 10.3102/1076998620958383 Article reuse guidelines: sagepub.com/journals-permissions © 2020 AERA. http://jebs.aera.net
Christian Aßmann、班贝格-纪尧姆·巴塞大学、斯坦福大学Michela Battaz、乌迪内大学Eli Ben Michael、加州大学伯克利分校Howard Bloom、人力资源示范研究公司(MDRC)Ulf Bockenholt、西北大学Daniel Bolt、威斯康星大学Eric Bradlow、宾夕法尼亚大学沃顿商学院Henry Braun,波士顿学院Katherine Castellano、教育测试服务机构Wendy Chan、宾夕法尼亚大学Yinghan Chen、内华达大学Reno Yinyin Chen、伊利诺伊大学厄巴纳-香槟分校Yunxiao Chen、伦敦政治经济学院Edison Choe、研究生管理招生委员会Steven Culpepper、,俄亥俄州立大学Jimmy de la Torre、香港大学Dries Debeer、鲁汶大学(KU Leuven)Peng Ding、加州大学伯克利分校Nianbo Dong、北卡罗来纳大学教堂山分校Fritz Drasgow、伊利诺伊大学厄巴纳-香槟分校Keelan Evanini、教育测试服务机构Jean-Paul Fox、特文特大学Mark Fredrickson、,密歇根大学Jorge Gonzaléz、智利天主教大学Simon Grund、莱布尼茨自然科学与数学研究所和基尔大学郭洪文、教育测试服务机构Gregory Hancock、马里兰大学Ben Hansen、密歇根大学Jeffrey Harring、马里兰大学Johannes Hartig,德国国际教育研究所(DIPF)Michael Harwell,明尼苏达大学Timothy Hayes,佛罗里达国际大学Yong He,ACT,股份有限公司Carolyn Hill,人力示范研究公司(MDRC)Minjeong Jeon,加州大学洛杉矶分校Paul Jewsbury,教育测试服务机构Booil Jo,斯坦福大学Harry Joe,不列颠哥伦比亚大学Matthew Johnson,教育测试服务机构George Karabatsos,伊利诺伊大学Chicago Luke Keele,宾夕法尼亚大学Augustin Kelava,Eberhard Karls University Tubingen Journal of Educational and Behavioral Statistics 2020,Vol.45,No.6,第771–773页DOI:10.3102/1076998620958383文章重用指南:sagepub.com/journals-permissions©2020 AERA。http://jebs.aera.net
{"title":"Acknowledgments","authors":"C. Aßmann, Yinyin Chen, Bryan Keller, Jee-Seon Kim, K. Kim","doi":"10.3102/1076998620958383","DOIUrl":"https://doi.org/10.3102/1076998620958383","url":null,"abstract":"Christian Aßmann, University of Bamberg Guillaume Basse, Stanford University Michela Battauz, University of Udine Eli Ben-Michael, University of CaliforniaBerkeley Howard Bloom, Manpower Demonstration Research Corporation (MDRC) Ulf Bockenholt, Northwestern University Daniel Bolt, University of Wisconsin Eric Bradlow, Wharton School of the University of Pennsylvania Henry Braun, Boston College Katherine Castellano, Educational Testing Service Wendy Chan, University of Pennsylvania Yinghan Chen, University of Nevada, Reno Yinyin Chen, University of Illinois at Urbana-Champaign Yunxiao Chen, London School of Economics and Political Science Edison Choe, Graduate Management Admission Council Steven Culpepper, University of Illinois at Urbana-Champaign Paul De Boeck, The Ohio State University Jimmy de la Torre, The University of Hong Kong Dries Debeer, University of Leuven (KU Leuven) Peng Ding, University of CaliforniaBerkeley Nianbo Dong, University of North Carolina at Chapel Hill Fritz Drasgow, University of Illinois at Urbana-Champaign Keelan Evanini, Educational Testing Service Jean-Paul Fox, University of Twente Mark Fredrickson, University of Michigan Jorge Gonzaléz, Pontificia Universidad Catolica de Chile Simon Grund, Leibniz-Institut fur die Padagogik der Naturwissenschaften und Mathematik an der Universitat Kiel Hongwen Guo, Educational Testing Service Gregory Hancock, University of MarylandCollege Park Ben Hansen, University of Michigan Jeffrey Harring, University of MarylandCollege Park Johannes Hartig, German Institute for International Educational Research (DIPF) Michael Harwell, University of Minnesota Timothy Hayes, Florida International University Yong He, ACT, Inc. Carolyn Hill, Manpower Demonstration Research Corporation (MDRC) Minjeong Jeon, University of CaliforniaLos Angeles Paul Jewsbury, Educational Testing Service Booil Jo, Stanford University Harry Joe, University of British Columbia Matthew Johnson, Educational Testing Service George Karabatsos, University of IllinoisChicago Luke Keele, University of Pennsylvania Augustin Kelava, Eberhard Karls University Tuebingen Journal of Educational and Behavioral Statistics 2020, Vol. 45, No. 6, pp. 771–773 DOI: 10.3102/1076998620958383 Article reuse guidelines: sagepub.com/journals-permissions © 2020 AERA. http://jebs.aera.net","PeriodicalId":48001,"journal":{"name":"Journal of Educational and Behavioral Statistics","volume":"45 1","pages":"771 - 773"},"PeriodicalIF":2.4,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43561935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Educational and Behavioral Statistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1