首页 > 最新文献

Educational and Psychological Measurement最新文献

英文 中文
The Accuracy of Bayesian Model Fit Indices in Selecting Among Multidimensional Item Response Theory Models. 贝叶斯模型拟合指标在多维项目反应理论模型选择中的准确性
IF 2.1 3区 心理学 Q1 Social Sciences Pub Date : 2024-04-01 Epub Date: 2023-05-25 DOI: 10.1177/00131644231165520
Ken A Fujimoto, Carl F Falk

Item response theory (IRT) models are often compared with respect to predictive performance to determine the dimensionality of rating scale data. However, such model comparisons could be biased toward nested-dimensionality IRT models (e.g., the bifactor model) when comparing those models with non-nested-dimensionality IRT models (e.g., a unidimensional or a between-item-dimensionality model). The reason is that, compared with non-nested-dimensionality models, nested-dimensionality models could have a greater propensity to fit data that do not represent a specific dimensional structure. However, it is unclear as to what degree model comparison results are biased toward nested-dimensionality IRT models when the data represent specific dimensional structures and when Bayesian estimation and model comparison indices are used. We conducted a simulation study to add clarity to this issue. We examined the accuracy of four Bayesian predictive performance indices at differentiating among non-nested- and nested-dimensionality IRT models. The deviance information criterion (DIC), a commonly used index to compare Bayesian models, was extremely biased toward nested-dimensionality IRT models, favoring them even when non-nested-dimensionality models were the correct models. The Pareto-smoothed importance sampling approximation of the leave-one-out cross-validation was the least biased, with the Watanabe information criterion and the log-predicted marginal likelihood closely following. The findings demonstrate that nested-dimensionality IRT models are not automatically favored when the data represent specific dimensional structures as long as an appropriate predictive performance index is used.

项目反应理论(IRT)模型通常与预测性能进行比较,以确定评分量表数据的维度。然而,当将这些模型与非嵌套维度的IRT模型(例如,一维或项目间维度模型)进行比较时,这种模型比较可能偏向于嵌套维度的IRT模型(例如双因子模型)。原因是,与非嵌套维度模型相比,嵌套维度模型可能更倾向于拟合不代表特定维度结构的数据。然而,当数据表示特定的维度结构时,以及当使用贝叶斯估计和模型比较指数时,尚不清楚模型比较结果在多大程度上偏向嵌套维度IRT模型。我们进行了一项模拟研究,以澄清这一问题。我们检验了四个贝叶斯预测性能指标在区分非嵌套维度和嵌套维度IRT模型方面的准确性。偏差信息准则(DIC)是比较贝叶斯模型的常用指标,它极倾向于嵌套维度的IRT模型,即使非嵌套维度模型是正确的模型,也有利于它们。留一交叉验证的Pareto平滑重要性抽样近似偏差最小,Watanabe信息准则和对数预测边际似然紧随其后。研究结果表明,只要使用适当的预测性能指数,当数据表示特定的维度结构时,嵌套维度IRT模型就不会自动受到青睐。
{"title":"The Accuracy of Bayesian Model Fit Indices in Selecting Among Multidimensional Item Response Theory Models.","authors":"Ken A Fujimoto, Carl F Falk","doi":"10.1177/00131644231165520","DOIUrl":"10.1177/00131644231165520","url":null,"abstract":"<p><p>Item response theory (IRT) models are often compared with respect to predictive performance to determine the dimensionality of rating scale data. However, such model comparisons could be biased toward nested-dimensionality IRT models (e.g., the bifactor model) when comparing those models with non-nested-dimensionality IRT models (e.g., a unidimensional or a between-item-dimensionality model). The reason is that, compared with non-nested-dimensionality models, nested-dimensionality models could have a greater propensity to fit data that do not represent a specific dimensional structure. However, it is unclear as to what degree model comparison results are biased toward nested-dimensionality IRT models when the data represent specific dimensional structures and when Bayesian estimation and model comparison indices are used. We conducted a simulation study to add clarity to this issue. We examined the accuracy of four Bayesian predictive performance indices at differentiating among non-nested- and nested-dimensionality IRT models. The deviance information criterion (DIC), a commonly used index to compare Bayesian models, was extremely biased toward nested-dimensionality IRT models, favoring them even when non-nested-dimensionality models were the correct models. The Pareto-smoothed importance sampling approximation of the leave-one-out cross-validation was the least biased, with the Watanabe information criterion and the log-predicted marginal likelihood closely following. The findings demonstrate that nested-dimensionality IRT models are not automatically favored when the data represent specific dimensional structures as long as an appropriate predictive performance index is used.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11185105/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48119563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dominance Analysis for Latent Variable Models: A Comparison of Methods With Categorical Indicators and Misspecified Models. 潜在变量模型的优势分析:分类指标和未指定模型方法的比较
IF 2.1 3区 心理学 Q1 Social Sciences Pub Date : 2024-04-01 Epub Date: 2023-04-28 DOI: 10.1177/00131644231171751
W Holmes Finch

Dominance analysis (DA) is a very useful tool for ordering independent variables in a regression model based on their relative importance in explaining variance in the dependent variable. This approach, which was originally described by Budescu, has recently been extended to use with structural equation models examining relationships among latent variables. Research demonstrated that this approach yields accurate results for latent variable models involving normally distributed indicator variables and correctly specified models. The purpose of the current simulation study was to compare the use of this DA approach to a method based on observed regression DA and DA when the latent variable model is estimated using two-stage least squares for latent variable models with categorical indicators and/or model misspecification. Results indicated that the DA approach for latent variable models can provide accurate ordering of the variables and correct hypothesis selection when indicators are categorical and models are misspecified. A discussion of implications from this study is provided.

优势分析(DA)是一种非常有用的工具,可以根据自变量在解释因变量方差中的相对重要性对回归模型中的自变量进行排序。这种方法最初由Budescu描述,最近被扩展到用于检查潜在变量之间关系的结构方程模型。研究表明,这种方法对涉及正态分布指标变量和正确指定模型的潜在变量模型产生了准确的结果。当前模拟研究的目的是将这种DA方法的使用与基于观测回归DA和DA的方法进行比较,当使用两阶段最小二乘法对具有分类指标和/或模型错误指定的潜在变量模型进行估计时。结果表明,当指标是分类的,模型是错误指定的时,潜在变量模型的DA方法可以提供变量的准确排序和正确的假设选择。对本研究的影响进行了讨论。
{"title":"Dominance Analysis for Latent Variable Models: A Comparison of Methods With Categorical Indicators and Misspecified Models.","authors":"W Holmes Finch","doi":"10.1177/00131644231171751","DOIUrl":"10.1177/00131644231171751","url":null,"abstract":"<p><p>Dominance analysis (DA) is a very useful tool for ordering independent variables in a regression model based on their relative importance in explaining variance in the dependent variable. This approach, which was originally described by Budescu, has recently been extended to use with structural equation models examining relationships among latent variables. Research demonstrated that this approach yields accurate results for latent variable models involving normally distributed indicator variables and correctly specified models. The purpose of the current simulation study was to compare the use of this DA approach to a method based on observed regression DA and DA when the latent variable model is estimated using two-stage least squares for latent variable models with categorical indicators and/or model misspecification. Results indicated that the DA approach for latent variable models can provide accurate ordering of the variables and correct hypothesis selection when indicators are categorical and models are misspecified. A discussion of implications from this study is provided.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11185102/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44118004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Trade-Off Between Factor Score Determinacy and the Preservation of Inter-Factor Correlations. 因子得分确定性与因子间相关性保持之间的权衡
IF 2.1 3区 心理学 Q1 Social Sciences Pub Date : 2024-04-01 Epub Date: 2023-04-29 DOI: 10.1177/00131644231171137
André Beauducel, Norbert Hilger, Tobias Kuhl

Regression factor score predictors have the maximum factor score determinacy, that is, the maximum correlation with the corresponding factor, but they do not have the same inter-correlations as the factors. As it might be useful to compute factor score predictors that have the same inter-correlations as the factors, correlation-preserving factor score predictors have been proposed. However, correlation-preserving factor score predictors have smaller correlations with the corresponding factors (factor score determinacy) than regression factor score predictors. Thus, higher factor score determinacy goes along with bias of the inter-correlations and unbiased inter-correlations go along with lower factor score determinacy. The aim of the present study was therefore to investigate the size and conditions of the trade-off between factor score determinacy and bias of inter-correlations by means of algebraic considerations and a simulation study. It turns out that under several conditions very small gains of factor score determinacy of the regression factor score predictor go along with a large bias of inter-correlations. Instead of using the regression factor score predictor by default, it is proposed to check whether substantial bias of inter-correlations can be avoided without substantial loss of factor score determinacy using a correlation-preserving factor score predictor. A syntax that allows to compute correlation-preserving factor score predictors from regression factor score predictors, and to compare factor score determinacy and inter-correlations of the factor score predictors is given in the Appendix.

回归因子得分预测因子具有最大因子得分确定性,即与相应因子的最大相关性,但它们与因子不具有相同的相关性。由于计算与因子具有相同相互相关性的因子得分预测因子可能是有用的,因此提出了保留相关性的因子分数预测因子。然而,保持相关性的因子得分预测因子与相应因子的相关性(因子得分的确定性)小于回归因子得分预测函数。因此,因子得分的确定性越高,相互关联的偏倚就越大,而无偏倚的相互关联就越低。因此,本研究的目的是通过代数考虑和模拟研究,研究因子得分确定性和相关性偏差之间的权衡大小和条件。结果表明,在几种条件下,回归因子得分预测因子的因子得分确定性的非常小的增益伴随着相关性的大偏差。建议使用保留相关性的因子得分预测器来检查是否可以在不显著损失因子得分确定性的情况下避免相互相关性的显著偏差,而不是默认使用回归因子得分预测因子。附录中给出了一种语法,该语法允许从回归因子得分预测因子中计算保持相关性的因子得分预测函数,并比较因子得分的确定性和因子得分预测的相互相关性。
{"title":"The Trade-Off Between Factor Score Determinacy and the Preservation of Inter-Factor Correlations.","authors":"André Beauducel, Norbert Hilger, Tobias Kuhl","doi":"10.1177/00131644231171137","DOIUrl":"10.1177/00131644231171137","url":null,"abstract":"<p><p>Regression factor score predictors have the maximum factor score determinacy, that is, the maximum correlation with the corresponding factor, but they do not have the same inter-correlations as the factors. As it might be useful to compute factor score predictors that have the same inter-correlations as the factors, correlation-preserving factor score predictors have been proposed. However, correlation-preserving factor score predictors have smaller correlations with the corresponding factors (factor score determinacy) than regression factor score predictors. Thus, higher factor score determinacy goes along with bias of the inter-correlations and unbiased inter-correlations go along with lower factor score determinacy. The aim of the present study was therefore to investigate the size and conditions of the trade-off between factor score determinacy and bias of inter-correlations by means of algebraic considerations and a simulation study. It turns out that under several conditions very small gains of factor score determinacy of the regression factor score predictor go along with a large bias of inter-correlations. Instead of using the regression factor score predictor by default, it is proposed to check whether substantial bias of inter-correlations can be avoided without substantial loss of factor score determinacy using a correlation-preserving factor score predictor. A syntax that allows to compute correlation-preserving factor score predictors from regression factor score predictors, and to compare factor score determinacy and inter-correlations of the factor score predictors is given in the Appendix.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.1,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11185104/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48879571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fused SDT/IRT Models for Mixed-Format Exams 混合格式考试的融合 SDT/IRT 模型
IF 2.7 3区 心理学 Q1 Social Sciences Pub Date : 2024-03-28 DOI: 10.1177/00131644241235333
Lawrence T. DeCarlo
A psychological framework for different types of items commonly used with mixed-format exams is proposed. A choice model based on signal detection theory (SDT) is used for multiple-choice (MC) items, whereas an item response theory (IRT) model is used for open-ended (OE) items. The SDT and IRT models are shown to share a common conceptualization in terms of latent states of “know/don’t know” at the examinee level. This in turn suggests a way to join or “fuse” the models—through the probability of knowing. A general model that fuses the SDT choice model, for MC items, with a generalized sequential logit model, for OE items, is introduced. Fitting SDT and IRT models simultaneously allows one to examine possible differences in psychological processes across the different types of items, to examine the effects of covariates in both models simultaneously, to allow for relations among the model parameters, and likely offers potential estimation benefits. The utility of the approach is illustrated with MC and OE items from large-scale international exams.
针对混合形式考试中常用的不同类型的题目,提出了一个心理学框架。基于信号检测理论(SDT)的选择模型适用于多项选择(MC)题目,而项目反应理论(IRT)模型则适用于开放式(OE)题目。结果表明,SDT 模型和 IRT 模型在被试者水平上的 "知道/不知道 "潜在状态方面具有共同的概念。这反过来又提出了一种通过 "知道 "的概率来连接或 "融合 "这两种模型的方法。本文介绍了一个通用模型,该模型融合了针对 MC 项目的 SDT 选择模型和针对 OE 项目的广义顺序 logit 模型。同时拟合 SDT 模型和 IRT 模型,可以考察不同类型项目的心理过程可能存在的差异,同时考察两个模型中协变量的影响,考虑模型参数之间的关系,并可能带来潜在的估算优势。我们用大型国际考试中的 MC 和 OE 项目来说明这种方法的实用性。
{"title":"Fused SDT/IRT Models for Mixed-Format Exams","authors":"Lawrence T. DeCarlo","doi":"10.1177/00131644241235333","DOIUrl":"https://doi.org/10.1177/00131644241235333","url":null,"abstract":"A psychological framework for different types of items commonly used with mixed-format exams is proposed. A choice model based on signal detection theory (SDT) is used for multiple-choice (MC) items, whereas an item response theory (IRT) model is used for open-ended (OE) items. The SDT and IRT models are shown to share a common conceptualization in terms of latent states of “know/don’t know” at the examinee level. This in turn suggests a way to join or “fuse” the models—through the probability of knowing. A general model that fuses the SDT choice model, for MC items, with a generalized sequential logit model, for OE items, is introduced. Fitting SDT and IRT models simultaneously allows one to examine possible differences in psychological processes across the different types of items, to examine the effects of covariates in both models simultaneously, to allow for relations among the model parameters, and likely offers potential estimation benefits. The utility of the approach is illustrated with MC and OE items from large-scale international exams.","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140322190","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Examining the Dynamic of Clustering Effects in Multilevel Designs: A Latent Variable Method Application 考察多层次设计中聚类效应的动态:潜变量法的应用
IF 2.7 3区 心理学 Q1 Social Sciences Pub Date : 2024-02-21 DOI: 10.1177/00131644241228602
Tenko Raykov, Ahmed Haddadi, Christine DiStefano, Mohammed Alqabbaa
This note is concerned with the study of temporal development in several indices reflecting clustering effects in multilevel designs that are frequently utilized in educational and behavioral research. A latent variable method-based approach is outlined, which can be used to point and interval estimate the growth or decline in important functions of level-specific variances in two-level and three-level settings. The procedure may also be employed for the purpose of examining stability over time in clustering effects. The method can be utilized with widely circulated latent variable modeling software, and is illustrated using empirical examples.
本说明主要研究在教育和行为研究中经常使用的多层次设计中反映聚类效应的几个指数的时间发展。本文概述了一种基于潜变量方法的方法,该方法可用于在两级和三级设置中对特定水平方差的重要函数的增长或下降进行点和区间估计。该方法还可用于研究聚类效应随时间变化的稳定性。该方法可与广泛使用的潜在变量建模软件结合使用,并通过经验实例加以说明。
{"title":"Examining the Dynamic of Clustering Effects in Multilevel Designs: A Latent Variable Method Application","authors":"Tenko Raykov, Ahmed Haddadi, Christine DiStefano, Mohammed Alqabbaa","doi":"10.1177/00131644241228602","DOIUrl":"https://doi.org/10.1177/00131644241228602","url":null,"abstract":"This note is concerned with the study of temporal development in several indices reflecting clustering effects in multilevel designs that are frequently utilized in educational and behavioral research. A latent variable method-based approach is outlined, which can be used to point and interval estimate the growth or decline in important functions of level-specific variances in two-level and three-level settings. The procedure may also be employed for the purpose of examining stability over time in clustering effects. The method can be utilized with widely circulated latent variable modeling software, and is illustrated using empirical examples.","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-02-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139954004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating Close Fit in Ordinal Factor Analysis Models With Multiply Imputed Data. 用多输入数据评价有序因子分析模型的紧密拟合
IF 2.7 3区 心理学 Q1 Social Sciences Pub Date : 2024-02-01 Epub Date: 2023-03-27 DOI: 10.1177/00131644231158854
Dexin Shi, Bo Zhang, Ren Liu, Zhehan Jiang

Multiple imputation (MI) is one of the recommended techniques for handling missing data in ordinal factor analysis models. However, methods for computing MI-based fit indices under ordinal factor analysis models have yet to be developed. In this short note, we introduced the methods of using the standardized root mean squared residual (SRMR) and the root mean square error of approximation (RMSEA) to assess the fit of ordinal factor analysis models with multiply imputed data. Specifically, we described the procedure for computing the MI-based sample estimates and constructing the confidence intervals. Simulation results showed that the proposed methods could yield sufficiently accurate point and interval estimates for both SRMR and RMSEA, especially in conditions with larger sample sizes, less missing data, more response categories, and higher degrees of misfit. Based on the findings, implications and recommendations were discussed.

多重插值(Multiple imputation, MI)是处理有序因子分析模型中缺失数据的推荐技术之一。然而,在有序因子分析模型下计算基于mi的拟合指数的方法尚未开发。在这篇简短的文章中,我们介绍了使用标准化均方根残差(SRMR)和近似均方根误差(RMSEA)的方法来评估多重输入数据的有序因子分析模型的拟合。具体来说,我们描述了计算基于mi的样本估计和构造置信区间的过程。仿真结果表明,本文提出的方法对SRMR和RMSEA都能产生足够精确的点和区间估计,特别是在样本量较大、缺失数据较少、响应类别较多、失配程度较高的情况下。根据调查结果,讨论了影响和建议。
{"title":"Evaluating Close Fit in Ordinal Factor Analysis Models With Multiply Imputed Data.","authors":"Dexin Shi, Bo Zhang, Ren Liu, Zhehan Jiang","doi":"10.1177/00131644231158854","DOIUrl":"10.1177/00131644231158854","url":null,"abstract":"<p><p>Multiple imputation (MI) is one of the recommended techniques for handling missing data in ordinal factor analysis models. However, methods for computing MI-based fit indices under ordinal factor analysis models have yet to be developed. In this short note, we introduced the methods of using the standardized root mean squared residual (SRMR) and the root mean square error of approximation (RMSEA) to assess the fit of ordinal factor analysis models with multiply imputed data. Specifically, we described the procedure for computing the MI-based sample estimates and constructing the confidence intervals. Simulation results showed that the proposed methods could yield sufficiently accurate point and interval estimates for both SRMR and RMSEA, especially in conditions with larger sample sizes, less missing data, more response categories, and higher degrees of misfit. Based on the findings, implications and recommendations were discussed.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10795567/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48895403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Are the Steps on Likert Scales Equidistant? Responses on Visual Analog Scales Allow Estimating Their Distances. 李克特量表上的台阶是等距的吗?在视觉模拟尺度上的反应允许估计它们的距离
IF 2.7 3区 心理学 Q1 Social Sciences Pub Date : 2024-02-01 Epub Date: 2023-04-04 DOI: 10.1177/00131644231164316
Miguel A García-Pérez

A recurring question regarding Likert items is whether the discrete steps that this response format allows represent constant increments along the underlying continuum. This question appears unsolvable because Likert responses carry no direct information to this effect. Yet, any item administered in Likert format can identically be administered with a continuous response format such as a visual analog scale (VAS) in which respondents mark a position along a continuous line. Then, the operating characteristics of the item would manifest under both VAS and Likert formats, although perhaps differently as captured by the continuous response model (CRM) and the graded response model (GRM) in item response theory. This article shows that CRM and GRM item parameters hold a formal relation that is mediated by the form in which the continuous dimension is partitioned into intervals to render the discrete Likert responses. Then, CRM and GRM characterizations of the items in a test administered with VAS and Likert formats allow estimating the boundaries of the partition that renders Likert responses for each item and, thus, the distance between consecutive steps. The validity of this approach is first documented via simulation studies. Subsequently, the same approach is used on public data from three personality scales with 12, eight, and six items, respectively. The results indicate the expected correspondence between VAS and Likert responses and reveal unequal distances between successive pairs of Likert steps that also vary greatly across items. Implications for the scoring of Likert items are discussed.

关于Likert项目的一个反复出现的问题是,这种响应格式允许的离散步骤是否表示沿着底层连续体的恒定增量。这个问题似乎无法解决,因为Likert的回答没有直接的信息。然而,以Likert格式管理的任何项目都可以相同地使用连续响应格式管理,例如视觉模拟量表(VAS),其中受访者沿着连续线标记位置。然后,项目的操作特征将在VAS和Likert格式下表现出来,尽管可能与项目响应理论中的连续响应模型(CRM)和分级响应模型(GRM)不同。本文表明,CRM和GRM项目参数具有一种形式关系,这种关系是由连续维度划分为区间以呈现离散Likert响应的形式介导的。然后,在使用VAS和Likert格式进行的测试中,项目的CRM和GRM特征允许估计呈现每个项目的Likert响应的分区的边界,从而估计连续步骤之间的距离。该方法的有效性首先通过模拟研究得到证明。随后,同样的方法被用于三个人格量表的公共数据,分别为12、8和6个项目。结果表明VAS和Likert反应之间的预期对应性,并揭示了连续的Likert步骤对之间的不相等距离,这些距离在不同项目之间也有很大差异。讨论了Likert项目评分的含义。
{"title":"Are the Steps on Likert Scales Equidistant? Responses on Visual Analog Scales Allow Estimating Their Distances.","authors":"Miguel A García-Pérez","doi":"10.1177/00131644231164316","DOIUrl":"10.1177/00131644231164316","url":null,"abstract":"<p><p>A recurring question regarding Likert items is whether the discrete steps that this response format allows represent constant increments along the underlying continuum. This question appears unsolvable because Likert responses carry no direct information to this effect. Yet, any item administered in Likert format can identically be administered with a continuous response format such as a visual analog scale (VAS) in which respondents mark a position along a continuous line. Then, the operating characteristics of the item would manifest under both VAS and Likert formats, although perhaps differently as captured by the continuous response model (CRM) and the graded response model (GRM) in item response theory. This article shows that CRM and GRM item parameters hold a formal relation that is mediated by the form in which the continuous dimension is partitioned into intervals to render the discrete Likert responses. Then, CRM and GRM characterizations of the items in a test administered with VAS and Likert formats allow estimating the boundaries of the partition that renders Likert responses for each item and, thus, the distance between consecutive steps. The validity of this approach is first documented via simulation studies. Subsequently, the same approach is used on public data from three personality scales with 12, eight, and six items, respectively. The results indicate the expected correspondence between VAS and Likert responses and reveal unequal distances between successive pairs of Likert steps that also vary greatly across items. Implications for the scoring of Likert items are discussed.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10795572/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48541417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Equating Oral Reading Fluency Scores: A Model-Based Approach. 一种基于模型的口语阅读流利度评分等值方法
IF 2.7 3区 心理学 Q1 Social Sciences Pub Date : 2024-02-01 Epub Date: 2023-01-05 DOI: 10.1177/00131644221148122
Yusuf Kara, Akihito Kamata, Xin Qiao, Cornelis J Potgieter, Joseph F T Nese

Words read correctly per minute (WCPM) is the reporting score metric in oral reading fluency (ORF) assessments, which is popularly utilized as part of curriculum-based measurements to screen at-risk readers and to monitor progress of students who receive interventions. Just like other types of assessments with multiple forms, equating would be necessary when WCPM scores are obtained from multiple ORF passages to be compared both between and within students. This article proposes a model-based approach for equating WCPM scores. A simulation study was conducted to evaluate the performance of the model-based equating approach along with some observed-score equating methods with external anchor test design.

每分钟正确阅读单词(WCPM)是口语阅读流利度(ORF)评估中的报告分数指标,它被广泛用作基于课程的测量的一部分,以筛选有风险的读者并监测接受干预的学生的进展。就像其他类型的多种形式的评估一样,当WCPM分数是从多个ORF段落中获得的,并在学生之间和学生内部进行比较时,将其等同起来是必要的。本文提出了一种基于模型的WCPM分数等值方法。进行了一项模拟研究,以评估基于模型的等值方法的性能,以及一些观察到的外部锚试验设计的分数等值方法。
{"title":"Equating Oral Reading Fluency Scores: A Model-Based Approach.","authors":"Yusuf Kara, Akihito Kamata, Xin Qiao, Cornelis J Potgieter, Joseph F T Nese","doi":"10.1177/00131644221148122","DOIUrl":"10.1177/00131644221148122","url":null,"abstract":"<p><p>Words read correctly per minute (WCPM) is the reporting score metric in oral reading fluency (ORF) assessments, which is popularly utilized as part of curriculum-based measurements to screen at-risk readers and to monitor progress of students who receive interventions. Just like other types of assessments with multiple forms, equating would be necessary when WCPM scores are obtained from multiple ORF passages to be compared both between and within students. This article proposes a model-based approach for equating WCPM scores. A simulation study was conducted to evaluate the performance of the model-based equating approach along with some observed-score equating methods with external anchor test design.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10795571/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41418799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial Neural Networks for Short-Form Development of Psychometric Tests: A Study on Synthetic Populations Using Autoencoders. 用于心理测量测试短期开发的人工神经网络:使用自动编码器对合成群体的研究
IF 2.7 3区 心理学 Q1 Social Sciences Pub Date : 2024-02-01 Epub Date: 2023-04-15 DOI: 10.1177/00131644231164363
Monica Casella, Pasquale Dolce, Michela Ponticorvo, Nicola Milano, Davide Marocco

Short-form development is an important topic in psychometric research, which requires researchers to face methodological choices at different steps. The statistical techniques traditionally used for shortening tests, which belong to the so-called exploratory model, make assumptions not always verified in psychological data. This article proposes a machine learning-based autonomous procedure for short-form development that combines explanatory and predictive techniques in an integrative approach. The study investigates the item-selection performance of two autoencoders: a particular type of artificial neural network that is comparable to principal component analysis. The procedure is tested on artificial data simulated from a factor-based population and is compared with existent computational approaches to develop short forms. Autoencoders require mild assumptions on data characteristics and provide a method to predict long-form items' responses from the short form. Indeed, results show that they can help the researcher to develop a short form by automatically selecting a subset of items that better reconstruct the original item's responses and that preserve the internal structure of the long-form.

简式发展是心理测量学研究中的一个重要课题,它要求研究人员在不同的步骤中面对方法论的选择。传统上用于缩短测试的统计技术属于所谓的探索性模型,其假设并不总是在心理数据中得到验证。本文提出了一种基于机器学习的简短开发自主程序,该程序将解释和预测技术结合在一起。该研究调查了两种自动编码器的项目选择性能:一种特殊类型的人工神经网络,可与主成分分析相媲美。该程序在基于因子的总体模拟的人工数据上进行了测试,并与现有的计算方法进行了比较,以开发简短的形式。自动编码器需要对数据特征进行温和的假设,并提供了一种从短格式中预测长格式项目响应的方法。事实上,研究结果表明,它们可以通过自动选择一个子集来帮助研究人员开发短格式,该子集可以更好地重建原始项目的响应,并保留长格式的内部结构。
{"title":"Artificial Neural Networks for Short-Form Development of Psychometric Tests: A Study on Synthetic Populations Using Autoencoders.","authors":"Monica Casella, Pasquale Dolce, Michela Ponticorvo, Nicola Milano, Davide Marocco","doi":"10.1177/00131644231164363","DOIUrl":"10.1177/00131644231164363","url":null,"abstract":"<p><p>Short-form development is an important topic in psychometric research, which requires researchers to face methodological choices at different steps. The statistical techniques traditionally used for shortening tests, which belong to the so-called exploratory model, make assumptions not always verified in psychological data. This article proposes a machine learning-based autonomous procedure for short-form development that combines explanatory and predictive techniques in an integrative approach. The study investigates the item-selection performance of two autoencoders: a particular type of artificial neural network that is comparable to principal component analysis. The procedure is tested on artificial data simulated from a factor-based population and is compared with existent computational approaches to develop short forms. Autoencoders require mild assumptions on data characteristics and provide a method to predict long-form items' responses from the short form. Indeed, results show that they can help the researcher to develop a short form by automatically selecting a subset of items that better reconstruct the original item's responses and that preserve the internal structure of the long-form.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10795568/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41603099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating Model Fit of Measurement Models in Confirmatory Factor Analysis. 验证因子分析中度量模型的模型拟合度评价
IF 2.7 3区 心理学 Q1 Social Sciences Pub Date : 2024-02-01 Epub Date: 2023-04-02 DOI: 10.1177/00131644231163813
David Goretzko, Karik Siemund, Philipp Sterner

Confirmatory factor analyses (CFA) are often used in psychological research when developing measurement models for psychological constructs. Evaluating CFA model fit can be quite challenging, as tests for exact model fit may focus on negligible deviances, while fit indices cannot be interpreted absolutely without specifying thresholds or cutoffs. In this study, we review how model fit in CFA is evaluated in psychological research using fit indices and compare the reported values with established cutoff rules. For this, we collected data on all CFA models in Psychological Assessment from the years 2015 to 2020 (NStudies=221). In addition, we reevaluate model fit with newly developed methods that derive fit index cutoffs that are tailored to the respective measurement model and the data characteristics at hand. The results of our review indicate that the model fit in many studies has to be seen critically, especially with regard to the usually imposed independent clusters constraints. In addition, many studies do not fully report all results that are necessary to re-evaluate model fit. We discuss these findings against new developments in model fit evaluation and methods for specification search.

在心理学研究中,当开发心理结构的测量模型时,经常使用证实性因素分析(CFA)。尽管评估CFA模型拟合可能非常具有挑战性,因为精确模型拟合的测试可能集中在可忽略的偏差上,而在没有指定阈值或截止值的情况下,拟合指数不能完全解释。在这项研究中,我们回顾了在心理学研究中如何使用拟合指数来评估CFA中的模型拟合,并将报告的值与既定的截断规则进行比较。为此,我们收集了2015年至2020年心理评估中所有CFA模型的数据[公式:见正文]。此外,我们使用新开发的方法重新评估模型拟合,这些方法导出了适合各自测量模型和手头数据特征的拟合指数截止值。我们的综述结果表明,必须严格看待模型在许多研究中的适用性,特别是在通常施加的独立集群约束方面。此外,许多研究并没有完全报告重新评估模型拟合所需的所有结果。我们针对模型拟合评估和规范搜索方法的新发展讨论了这些发现。
{"title":"Evaluating Model Fit of Measurement Models in Confirmatory Factor Analysis.","authors":"David Goretzko, Karik Siemund, Philipp Sterner","doi":"10.1177/00131644231163813","DOIUrl":"10.1177/00131644231163813","url":null,"abstract":"<p><p>Confirmatory factor analyses (CFA) are often used in psychological research when developing measurement models for psychological constructs. Evaluating CFA model fit can be quite challenging, as tests for exact model fit may focus on negligible deviances, while fit indices cannot be interpreted absolutely without specifying thresholds or cutoffs. In this study, we review how model fit in CFA is evaluated in psychological research using fit indices and compare the reported values with established cutoff rules. For this, we collected data on all CFA models in <i>Psychological Assessment</i> from the years 2015 to 2020 <math><mrow><mo>(</mo><msub><mrow><mi>N</mi></mrow><mrow><mi>Studies</mi></mrow></msub><mo>=</mo><mn>221</mn><mo>)</mo></mrow></math>. In addition, we reevaluate model fit with newly developed methods that derive fit index cutoffs that are tailored to the respective measurement model and the data characteristics at hand. The results of our review indicate that the model fit in many studies has to be seen critically, especially with regard to the usually imposed independent clusters constraints. In addition, many studies do not fully report all results that are necessary to re-evaluate model fit. We discuss these findings against new developments in model fit evaluation and methods for specification search.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10795573/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43250321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Educational and Psychological Measurement
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1