首页 > 最新文献

British Journal of Mathematical & Statistical Psychology最新文献

英文 中文
A comparison of different measures of the proportion of explained variance in multiply imputed data sets 比较多重估算数据集解释方差比例的不同测量方法
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-04-05 DOI: 10.1111/bmsp.12344
Joost R. van Ginkel, Julian D. Karch
<p>The proportion of explained variance is an important statistic in multiple regression for determining how well the outcome variable is predicted by the predictors. Earlier research on 20 different estimators for the proportion of explained variance, including the exact Olkin–Pratt estimator and the Ezekiel estimator, showed that the exact Olkin–Pratt estimator produced unbiased estimates, and was recommended as a default estimator. In the current study, the same 20 estimators were studied in incomplete data, with missing data being treated using multiple imputation. In earlier research on the proportion of explained variance in multiply imputed data sets, an estimator called <span></span><math> <semantics> <mrow> <msubsup> <mover> <mi>R</mi> <mo>̂</mo> </mover> <mi>SP</mi> <mn>2</mn> </msubsup> </mrow> </semantics></math> was shown to be the preferred pooled estimator for regular <span></span><math> <semantics> <mrow> <msup> <mi>R</mi> <mn>2</mn> </msup> </mrow> </semantics></math>. For each of the 20 estimators in the current study, two pooled estimators were proposed: one where the estimator was the average across imputed data sets, and one where <span></span><math> <semantics> <mrow> <msubsup> <mover> <mi>R</mi> <mo>̂</mo> </mover> <mi>SP</mi> <mn>2</mn> </msubsup> </mrow> </semantics></math> was used as input for the calculation of the specific estimator. Simulations showed that estimates based on <span></span><math> <semantics> <mrow> <msubsup> <mover> <mi>R</mi> <mo>̂</mo> </mover> <mi>SP</mi> <mn>2</mn> </msubsup> </mrow> </semantics></math> performed best regarding bias and accuracy, and that the Ezekiel estimator was generally the least biased. However, none of the estimators were unbiased at all times, including the exact Olkin–Pratt estimator based on <span></span><math> <semantics> <mrow> <msubsup> <mover> <mi>R</mi> <mo>̂</mo> </mover> <mi>SP
解释方差比例是多元回归中的一个重要统计量,用于确定预测变量对结果变量的预测程度。早先对 20 种不同的解释方差比例估计器(包括精确的 Olkin-Pratt 估计器和 Ezekiel 估计器)进行的研究表明,精确的 Olkin-Pratt 估计器能产生无偏估计,并被推荐为默认估计器。在本研究中,同样的 20 个估计器在不完整数据中进行了研究,缺失数据采用多重估算法处理。在早先对多重归因数据集解释方差比例的研究中,一个名为的估计器被证明是常规的首选集合估计器。对于当前研究中的 20 个估计器,分别提出了两个集合估计器:一个估计器是各归因数据集的平均值,另一个估计器是计算特定估计器的输入值。模拟结果表明,在偏差和准确性方面,以 Ezekiel 为基础的估计值表现最佳,而 Ezekiel 估计值通常偏差最小。然而,没有一个估计器在任何时候都是无偏的,包括基于 的精确奥尔金-普拉特估计器。
{"title":"A comparison of different measures of the proportion of explained variance in multiply imputed data sets","authors":"Joost R. van Ginkel,&nbsp;Julian D. Karch","doi":"10.1111/bmsp.12344","DOIUrl":"10.1111/bmsp.12344","url":null,"abstract":"&lt;p&gt;The proportion of explained variance is an important statistic in multiple regression for determining how well the outcome variable is predicted by the predictors. Earlier research on 20 different estimators for the proportion of explained variance, including the exact Olkin–Pratt estimator and the Ezekiel estimator, showed that the exact Olkin–Pratt estimator produced unbiased estimates, and was recommended as a default estimator. In the current study, the same 20 estimators were studied in incomplete data, with missing data being treated using multiple imputation. In earlier research on the proportion of explained variance in multiply imputed data sets, an estimator called &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;msubsup&gt;\u0000 &lt;mover&gt;\u0000 &lt;mi&gt;R&lt;/mi&gt;\u0000 &lt;mo&gt;̂&lt;/mo&gt;\u0000 &lt;/mover&gt;\u0000 &lt;mi&gt;SP&lt;/mi&gt;\u0000 &lt;mn&gt;2&lt;/mn&gt;\u0000 &lt;/msubsup&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt; was shown to be the preferred pooled estimator for regular &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;msup&gt;\u0000 &lt;mi&gt;R&lt;/mi&gt;\u0000 &lt;mn&gt;2&lt;/mn&gt;\u0000 &lt;/msup&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt;. For each of the 20 estimators in the current study, two pooled estimators were proposed: one where the estimator was the average across imputed data sets, and one where &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;msubsup&gt;\u0000 &lt;mover&gt;\u0000 &lt;mi&gt;R&lt;/mi&gt;\u0000 &lt;mo&gt;̂&lt;/mo&gt;\u0000 &lt;/mover&gt;\u0000 &lt;mi&gt;SP&lt;/mi&gt;\u0000 &lt;mn&gt;2&lt;/mn&gt;\u0000 &lt;/msubsup&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt; was used as input for the calculation of the specific estimator. Simulations showed that estimates based on &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;msubsup&gt;\u0000 &lt;mover&gt;\u0000 &lt;mi&gt;R&lt;/mi&gt;\u0000 &lt;mo&gt;̂&lt;/mo&gt;\u0000 &lt;/mover&gt;\u0000 &lt;mi&gt;SP&lt;/mi&gt;\u0000 &lt;mn&gt;2&lt;/mn&gt;\u0000 &lt;/msubsup&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt; performed best regarding bias and accuracy, and that the Ezekiel estimator was generally the least biased. However, none of the estimators were unbiased at all times, including the exact Olkin–Pratt estimator based on &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;msubsup&gt;\u0000 &lt;mover&gt;\u0000 &lt;mi&gt;R&lt;/mi&gt;\u0000 &lt;mo&gt;̂&lt;/mo&gt;\u0000 &lt;/mover&gt;\u0000 &lt;mi&gt;SP","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"672-693"},"PeriodicalIF":1.8,"publicationDate":"2024-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12344","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140592929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A two-step item bank calibration strategy based on 1-bit matrix completion for small-scale computerized adaptive testing 基于 1 位矩阵补全的两步式项目库校准策略,适用于小规模计算机自适应测试
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-04-04 DOI: 10.1111/bmsp.12340
Yawei Shen, Shiyu Wang, Houping Xiao

Computerized adaptive testing (CAT) is a widely embraced approach for delivering personalized educational assessments, tailoring each test to the real-time performance of individual examinees. Despite its potential advantages, CAT�s application in small-scale assessments has been limited due to the complexities associated with calibrating the item bank using sparse response data and small sample sizes. This study addresses these challenges by developing a two-step item bank calibration strategy that leverages the 1-bit matrix completion method in conjunction with two distinct incomplete pretesting designs. We introduce two novel 1-bit matrix completion-based imputation methods specifically designed to tackle the issues associated with item calibration in the presence of sparse response data and limited sample sizes. To demonstrate the effectiveness of these approaches, we conduct a comparative assessment against several established item parameter estimation methods capable of handling missing data. This evaluation is carried out through two sets of simulation studies, each featuring different pretesting designs, item bank structures, and sample sizes. Furthermore, we illustrate the practical application of the methods investigated, using empirical data collected from small-scale assessments.

计算机自适应测试(CAT)是一种被广泛接受的提供个性化教育评估的方法,它可以根据每个考生的实时表现量身定制每项测试。尽管CAT具有潜在的优势,但由于使用稀少的反应数据和小样本量校准题库的复杂性,它在小规模评估中的应用一直受到限制。为了应对这些挑战,本研究开发了一种分两步进行的项目库校准策略,利用 1 位矩阵补全法,结合两种不同的不完全前测设计,对项目库进行校准。我们介绍了两种新颖的基于 1 位矩阵补全的估算方法,专门用于解决稀疏响应数据和有限样本量情况下与项目校准相关的问题。为了证明这些方法的有效性,我们与几种能够处理缺失数据的成熟项目参数估计方法进行了比较评估。这项评估是通过两组模拟研究进行的,每组模拟研究都采用了不同的前测设计、题目库结构和样本量。此外,我们还利用从小规模评估中收集到的经验数据,说明了所研究方法的实际应用情况。
{"title":"A two-step item bank calibration strategy based on 1-bit matrix completion for small-scale computerized adaptive testing","authors":"Yawei Shen,&nbsp;Shiyu Wang,&nbsp;Houping Xiao","doi":"10.1111/bmsp.12340","DOIUrl":"10.1111/bmsp.12340","url":null,"abstract":"<p>Computerized adaptive testing (CAT) is a widely embraced approach for delivering personalized educational assessments, tailoring each test to the real-time performance of individual examinees. Despite its potential advantages, CAT�s application in small-scale assessments has been limited due to the complexities associated with calibrating the item bank using sparse response data and small sample sizes. This study addresses these challenges by developing a two-step item bank calibration strategy that leverages the 1-bit matrix completion method in conjunction with two distinct incomplete pretesting designs. We introduce two novel 1-bit matrix completion-based imputation methods specifically designed to tackle the issues associated with item calibration in the presence of sparse response data and limited sample sizes. To demonstrate the effectiveness of these approaches, we conduct a comparative assessment against several established item parameter estimation methods capable of handling missing data. This evaluation is carried out through two sets of simulation studies, each featuring different pretesting designs, item bank structures, and sample sizes. Furthermore, we illustrate the practical application of the methods investigated, using empirical data collected from small-scale assessments.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"553-612"},"PeriodicalIF":1.8,"publicationDate":"2024-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12340","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140592809","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sample size determination for interval estimation of the prevalence of a sensitive attribute under non-randomized response models 在非随机反应模型下,确定敏感属性流行率区间估计的样本量。
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-02-26 DOI: 10.1111/bmsp.12338
Shi-Fang Qiu, Jie Lei, Wai-Yin Poon, Man-Lai Tang, Ricky S. Wong, Ji-Ran Tao

A sufficient number of participants should be included to adequately address the research interest in the surveys with sensitive questions. In this paper, sample size formulas/iterative algorithms are developed from the perspective of controlling the confidence interval width of the prevalence of a sensitive attribute under four non-randomized response models: the crosswise model, parallel model, Poisson item count technique model and negative binomial item count technique model. In contrast to the conventional approach for sample size determination, our sample size formulas/algorithms explicitly incorporate an assurance probability of controlling the width of a confidence interval within the pre-specified range. The performance of the proposed methods is evaluated with respect to the empirical coverage probability, empirical assurance probability and confidence width. Simulation results show that all formulas/algorithms are effective and hence are recommended for practical applications. A real example is used to illustrate the proposed methods.

在涉及敏感问题的调查中,应纳入足够数量的参与者,以充分满足研究兴趣。本文从控制四种非随机响应模型(交叉模型、平行模型、泊松项目计数技术模型和负二项项目计数技术模型)下敏感属性流行率置信区间宽度的角度出发,建立了样本量计算公式/迭代算法。与确定样本容量的传统方法不同,我们的样本容量公式/算法明确包含了将置信区间宽度控制在预先指定范围内的保证概率。我们根据经验覆盖概率、经验保证概率和置信区间宽度对所提方法的性能进行了评估。仿真结果表明,所有公式/算法都是有效的,因此建议实际应用。一个真实的例子用于说明所提出的方法。
{"title":"Sample size determination for interval estimation of the prevalence of a sensitive attribute under non-randomized response models","authors":"Shi-Fang Qiu,&nbsp;Jie Lei,&nbsp;Wai-Yin Poon,&nbsp;Man-Lai Tang,&nbsp;Ricky S. Wong,&nbsp;Ji-Ran Tao","doi":"10.1111/bmsp.12338","DOIUrl":"10.1111/bmsp.12338","url":null,"abstract":"<p>A sufficient number of participants should be included to adequately address the research interest in the surveys with sensitive questions. In this paper, sample size formulas/iterative algorithms are developed from the perspective of controlling the confidence interval width of the prevalence of a sensitive attribute under four non-randomized response models: the crosswise model, parallel model, Poisson item count technique model and negative binomial item count technique model. In contrast to the conventional approach for sample size determination, our sample size formulas/algorithms explicitly incorporate an assurance probability of controlling the width of a confidence interval within the pre-specified range. The performance of the proposed methods is evaluated with respect to the empirical coverage probability, empirical assurance probability and confidence width. Simulation results show that all formulas/algorithms are effective and hence are recommended for practical applications. A real example is used to illustrate the proposed methods.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"508-531"},"PeriodicalIF":1.8,"publicationDate":"2024-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139974734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Assessment of fit of the time-varying dynamic partial credit model using the posterior predictive model checking method 使用后验预测模型检查法评估时变动态部分信贷模型的拟合度。
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-02-21 DOI: 10.1111/bmsp.12339
Sebastian Castro-Alvarez, Sandip Sinharay, Laura F. Bringmann, Rob R. Meijer, Jorge N. Tendeiro

Several new models based on item response theory have recently been suggested to analyse intensive longitudinal data. One of these new models is the time-varying dynamic partial credit model (TV-DPCM; Castro-Alvarez et al., Multivariate Behavioral Research, 2023, 1), which is a combination of the partial credit model and the time-varying autoregressive model. The model allows the study of the psychometric properties of the items and the modelling of nonlinear trends at the latent state level. However, there is a severe lack of tools to assess the fit of the TV-DPCM. In this paper, we propose and develop several test statistics and discrepancy measures based on the posterior predictive model checking (PPMC) method (PPMC; Rubin, The Annals of Statistics, 1984, 12, 1151) to assess the fit of the TV-DPCM. Simulated and empirical data are used to study the performance of and illustrate the effectiveness of the PPMC method.

最近,有人提出了几种基于项目反应理论的新模型来分析密集的纵向数据。其中一个新模型是时变动态部分学分模型(TV-DPCM;Castro-Alvarez 等人,《多变量行为研究》,2023 年第 1 期),它是部分学分模型和时变自回归模型的结合。该模型可以研究项目的心理测量特性,并在潜态水平上建立非线性趋势模型。然而,目前严重缺乏评估 TV-DPCM 拟合度的工具。在本文中,我们基于后验预测模型检查(PPMC)方法(PPMC; Rubin, The Annals of Statistics, 1984, 12, 1151)提出并开发了几种测试统计量和差异测量方法,用于评估 TV-DPCM 的拟合度。模拟数据和经验数据用于研究 PPMC 方法的性能并说明其有效性。
{"title":"Assessment of fit of the time-varying dynamic partial credit model using the posterior predictive model checking method","authors":"Sebastian Castro-Alvarez,&nbsp;Sandip Sinharay,&nbsp;Laura F. Bringmann,&nbsp;Rob R. Meijer,&nbsp;Jorge N. Tendeiro","doi":"10.1111/bmsp.12339","DOIUrl":"10.1111/bmsp.12339","url":null,"abstract":"<p>Several new models based on item response theory have recently been suggested to analyse intensive longitudinal data. One of these new models is the time-varying dynamic partial credit model (TV-DPCM; Castro-Alvarez et al., <i>Multivariate Behavioral Research</i>, 2023, 1), which is a combination of the partial credit model and the time-varying autoregressive model. The model allows the study of the psychometric properties of the items and the modelling of nonlinear trends at the latent state level. However, there is a severe lack of tools to assess the fit of the TV-DPCM. In this paper, we propose and develop several test statistics and discrepancy measures based on the posterior predictive model checking (PPMC) method (PPMC; Rubin, <i>The Annals of Statistics</i>, 1984, 12, 1151) to assess the fit of the TV-DPCM. Simulated and empirical data are used to study the performance of and illustrate the effectiveness of the PPMC method.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"532-552"},"PeriodicalIF":1.8,"publicationDate":"2024-02-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139914100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
When and how to use set-exploratory structural equation modelling to test structural models: A tutorial using the R package lavaan 何时以及如何使用集合探索式结构方程模型来检验结构模型:使用 R 软件包 lavaan 的教程。
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-02-15 DOI: 10.1111/bmsp.12336
Herb Marsh, Abdullah Alamer

Exploratory structural equation modelling (ESEM) is an alternative to the well-known method of confirmatory factor analysis (CFA). ESEM is mainly used to assess the quality of measurement models of common factors but can be efficiently extended to test structural models. However, ESEM may not be the best option in some model specifications, especially when structural models are involved, because the full flexibility of ESEM could result in technical difficulties in model estimation. Thus, set-ESEM was developed to accommodate the balance between full-ESEM and CFA. In the present paper, we show examples where set-ESEM should be used rather than full-ESEM. Rather than relying on a simulation study, we provide two applied examples using real data that are included in the OSF repository. Additionally, we provide the code needed to run set-ESEM in the free R package lavaan to make the paper practical. Set-ESEM structural models outperform their CFA-based counterparts in terms of goodness of fit and realistic factor correlation, and hence path coefficients in the two empirical examples. In several instances, effects that were non-significant (i.e., attenuated) in the CFA-based structural model become larger and significant in the set-ESEM structural model, suggesting that set-ESEM models may generate more accurate model parameters and, hence, lower Type II error rate.

探索性结构方程模型(ESEM)是著名的确证因素分析(CFA)方法的替代方法。ESEM 主要用于评估常见因子测量模型的质量,但也可以有效地扩展到测试结构模型。然而,ESEM 在某些模型规格中可能不是最佳选择,尤其是涉及结构模型时,因为 ESEM 的充分灵活性可能会导致模型估计中的技术困难。因此,为了兼顾完全 ESEM 和 CFA,我们开发了集合 ESEM。在本文中,我们将举例说明在哪些情况下应使用集合-ESEM,而不是完全-ESEM。我们没有依赖模拟研究,而是使用 OSF 存储库中的真实数据提供了两个应用实例。此外,我们还在免费的 R 软件包 lavaan 中提供了运行 Set-ESEM 所需的代码,从而使本文更加实用。在拟合优度和现实因子相关性方面,集合-ESEM 结构模型优于基于 CFA 的结构模型,因此在两个实证例子中的路径系数也优于基于 CFA 的结构模型。有几次,在基于 CFA 的结构模型中不显著(即衰减)的效应在集合-ESEM 结构模型中变得更大和显著,这表明集合-ESEM 模型可能会生成更准确的模型参数,从而降低 II 类错误率。
{"title":"When and how to use set-exploratory structural equation modelling to test structural models: A tutorial using the R package lavaan","authors":"Herb Marsh,&nbsp;Abdullah Alamer","doi":"10.1111/bmsp.12336","DOIUrl":"10.1111/bmsp.12336","url":null,"abstract":"<p>Exploratory structural equation modelling (ESEM) is an alternative to the well-known method of confirmatory factor analysis (CFA). ESEM is mainly used to assess the quality of measurement models of common factors but can be efficiently extended to test structural models. However, ESEM may not be the best option in some model specifications, especially when structural models are involved, because the full flexibility of ESEM could result in technical difficulties in model estimation. Thus, set-ESEM was developed to accommodate the balance between full-ESEM and CFA. In the present paper, we show examples where set-ESEM should be used rather than full-ESEM. Rather than relying on a simulation study, we provide two applied examples using real data that are included in the OSF repository. Additionally, we provide the code needed to run set-ESEM in the free R package <i>lavaan</i> to make the paper practical. Set-ESEM structural models outperform their CFA-based counterparts in terms of goodness of fit and realistic factor correlation, and hence path coefficients in the two empirical examples. In several instances, effects that were non-significant (i.e., attenuated) in the CFA-based structural model become larger and significant in the set-ESEM structural model, suggesting that set-ESEM models may generate more accurate model parameters and, hence, lower Type II error rate.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"459-476"},"PeriodicalIF":1.8,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12336","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139742778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast estimation of generalized linear latent variable models for performance and process data with ordinal, continuous, and count observed variables 快速估计具有顺序、连续和计数观测变量的性能和过程数据的广义线性潜变量模型。
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-02-12 DOI: 10.1111/bmsp.12337
Maoxin Zhang, Björn Andersson, Shaobo Jin

Different data types often occur in psychological and educational measurement such as computer-based assessments that record performance and process data (e.g., response times and the number of actions). Modelling such data requires specific models for each data type and accommodating complex dependencies between multiple variables. Generalized linear latent variable models are suitable for modelling mixed data simultaneously, but estimation can be computationally demanding. A fast solution is to use Laplace approximations, but existing implementations of joint modelling of mixed data types are limited to ordinal and continuous data. To address this limitation, we derive an efficient estimation method that uses first- or second-order Laplace approximations to simultaneously model ordinal data, continuous data, and count data. We illustrate the approach with an example and conduct simulations to evaluate the performance of the method in terms of estimation efficiency, convergence, and parameter recovery. The results suggest that the second-order Laplace approximation achieves a higher convergence rate and produces accurate yet fast parameter estimates compared to the first-order Laplace approximation, while the time cost increases with higher model complexity. Additionally, models that consider the dependence of variables from the same stimulus fit the empirical data substantially better than models that disregarded the dependence.

在心理和教育测量中经常会出现不同的数据类型,如记录表现和过程数据(如反应时间和操作次数)的基于计算机的评估。对这类数据建模需要针对每种数据类型建立特定的模型,并适应多个变量之间复杂的依赖关系。广义线性潜变量模型适用于同时对混合数据建模,但估算需要大量计算。快速的解决方案是使用拉普拉斯近似,但现有的混合数据类型联合建模方法仅限于序数和连续数据。为了解决这一局限性,我们推导出一种高效的估计方法,利用一阶或二阶拉普拉斯近似同时对序数数据、连续数据和计数数据建模。我们以实例说明了该方法,并进行了模拟,以评估该方法在估计效率、收敛性和参数恢复方面的性能。结果表明,与一阶拉普拉斯近似法相比,二阶拉普拉斯近似法能达到更高的收敛速度,并能产生准确而快速的参数估计,而时间成本会随着模型复杂度的提高而增加。此外,考虑同一刺激变量依赖性的模型比忽略依赖性的模型更符合经验数据。
{"title":"Fast estimation of generalized linear latent variable models for performance and process data with ordinal, continuous, and count observed variables","authors":"Maoxin Zhang,&nbsp;Björn Andersson,&nbsp;Shaobo Jin","doi":"10.1111/bmsp.12337","DOIUrl":"10.1111/bmsp.12337","url":null,"abstract":"<p>Different data types often occur in psychological and educational measurement such as computer-based assessments that record performance and process data (e.g., response times and the number of actions). Modelling such data requires specific models for each data type and accommodating complex dependencies between multiple variables. Generalized linear latent variable models are suitable for modelling mixed data simultaneously, but estimation can be computationally demanding. A fast solution is to use Laplace approximations, but existing implementations of joint modelling of mixed data types are limited to ordinal and continuous data. To address this limitation, we derive an efficient estimation method that uses first- or second-order Laplace approximations to simultaneously model ordinal data, continuous data, and count data. We illustrate the approach with an example and conduct simulations to evaluate the performance of the method in terms of estimation efficiency, convergence, and parameter recovery. The results suggest that the second-order Laplace approximation achieves a higher convergence rate and produces accurate yet fast parameter estimates compared to the first-order Laplace approximation, while the time cost increases with higher model complexity. Additionally, models that consider the dependence of variables from the same stimulus fit the empirical data substantially better than models that disregarded the dependence.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"477-507"},"PeriodicalIF":1.8,"publicationDate":"2024-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12337","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139725087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Constructing tests for skill assessment with competence-based test development 利用基于能力的测试开发构建技能评估测试
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-02-02 DOI: 10.1111/bmsp.12335
Pasquale Anselmi, Jürgen Heller, Luca Stefanutti, Egidio Robusto

Competence-based test development is a recent and innovative method for the construction of tests that are as informative as possible about the competence state (the set of skills an individual has available) underlying the observed item responses. It finds application in different contexts, including the development of tests from scratch, and the improvement or shortening of existing tests. Given a fixed collection of competence states existing in a population of individuals and a fixed collection of competencies (each of which being the subset of skills that allow for solving an item), the competency deletion procedure results in tests that differ from each other in the competencies but are all equally informative about individuals' competence states. This work introduces a streamlined version of the competency deletion procedure that considers information necessary for test construction only, illustrates a straightforward way to incorporate test developer preferences about competencies into the test construction process, and evaluates the performance of the resulting tests in uncovering the competence states from the observed item responses.

以能力为基础的测验开发是一种最新的创新方法,它所设计的测验能尽可能地反映所观察到的项目反应背后的能力状态(一个人所具备的一系列技能)。它适用于不同的情况,包括从零开始开发测验,以及改进或缩短现有的测验。在个人群体中存在固定的能力状态集合和固定的能力集合(每个能力集合都是能够解决一个题目的技能子集)的情况下,能力删除程序所产生的测验在能力方面彼此不同,但在个人能力状态方面却具有相同的信息量。这项工作介绍了一种简化版的能力删除程序,它只考虑了构建测验所需的信息,说明了一种将测验开发者对能力的偏好纳入测验构建过程的直接方法,并评估了由此产生的测验在从观察到的项目反应中揭示能力状态方面的性能。
{"title":"Constructing tests for skill assessment with competence-based test development","authors":"Pasquale Anselmi,&nbsp;Jürgen Heller,&nbsp;Luca Stefanutti,&nbsp;Egidio Robusto","doi":"10.1111/bmsp.12335","DOIUrl":"10.1111/bmsp.12335","url":null,"abstract":"<p>Competence-based test development is a recent and innovative method for the construction of tests that are as informative as possible about the competence state (the set of skills an individual has available) underlying the observed item responses. It finds application in different contexts, including the development of tests from scratch, and the improvement or shortening of existing tests. Given a fixed collection of competence states existing in a population of individuals and a fixed collection of competencies (each of which being the subset of skills that allow for solving an item), the competency deletion procedure results in tests that differ from each other in the competencies but are all equally informative about individuals' competence states. This work introduces a streamlined version of the competency deletion procedure that considers information necessary for test construction only, illustrates a straightforward way to incorporate test developer preferences about competencies into the test construction process, and evaluates the performance of the resulting tests in uncovering the competence states from the observed item responses.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"429-458"},"PeriodicalIF":1.8,"publicationDate":"2024-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139661345","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identifiability and estimability of Bayesian linear and nonlinear crossed random effects models 贝叶斯线性和非线性交叉随机效应模型的可识别性和可估算性。
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-01-24 DOI: 10.1111/bmsp.12334
Corissa T. Rohloff, Nidhi Kohli, Eric F. Lock

Crossed random effects models (CREMs) are particularly useful in longitudinal data applications because they allow researchers to account for the impact of dynamic group membership on individual outcomes. However, no research has determined what data conditions need to be met to sufficiently identify these models, especially the group effects, in a longitudinal context. This is a significant gap in the current literature as future applications to real data may need to consider these conditions to yield accurate and precise model parameter estimates, specifically for the group effects on individual outcomes. Furthermore, there are no existing CREMs that can model intrinsically nonlinear growth. The goals of this study are to develop a Bayesian piecewise CREM to model intrinsically nonlinear growth and evaluate what data conditions are necessary to empirically identify both intrinsically linear and nonlinear longitudinal CREMs. This study includes an applied example that utilizes the piecewise CREM with real data and three simulation studies to assess the data conditions necessary to estimate linear, quadratic, and piecewise CREMs. Results show that the number of repeated measurements collected on groups impacts the ability to recover the group effects. Additionally, functional form complexity impacted data collection requirements for estimating longitudinal CREMs.

交叉随机效应模型(CREMs)在纵向数据应用中特别有用,因为它们允许研究人员考虑动态群体成员身份对个体结果的影响。然而,目前还没有研究确定在纵向背景下需要满足哪些数据条件才能充分识别这些模型,尤其是群体效应。这是目前文献中的一个重要空白,因为未来在真实数据中的应用可能需要考虑这些条件,以获得准确和精确的模型参数估计,特别是群体效应对个体结果的影响。此外,还没有现有的 CREM 可以对内在非线性增长进行建模。本研究的目标是开发一种贝叶斯片断式 CREM,以模拟内在非线性增长,并评估哪些数据条件是通过经验识别内在线性和非线性纵向 CREM 所必需的。本研究包括一个利用真实数据利用片断 CREM 的应用实例和三项模拟研究,以评估估计线性、二次和片断 CREM 所需的数据条件。结果表明,收集到的群体重复测量次数会影响恢复群体效应的能力。此外,函数形式的复杂性也影响了估计纵向 CREM 的数据收集要求。
{"title":"Identifiability and estimability of Bayesian linear and nonlinear crossed random effects models","authors":"Corissa T. Rohloff,&nbsp;Nidhi Kohli,&nbsp;Eric F. Lock","doi":"10.1111/bmsp.12334","DOIUrl":"10.1111/bmsp.12334","url":null,"abstract":"<p>Crossed random effects models (CREMs) are particularly useful in longitudinal data applications because they allow researchers to account for the impact of dynamic group membership on individual outcomes. However, no research has determined what data conditions need to be met to sufficiently identify these models, especially the group effects, in a longitudinal context. This is a significant gap in the current literature as future applications to real data may need to consider these conditions to yield accurate and precise model parameter estimates, specifically for the group effects on individual outcomes. Furthermore, there are no existing CREMs that can model intrinsically nonlinear growth. The goals of this study are to develop a Bayesian piecewise CREM to model intrinsically nonlinear growth and evaluate what data conditions are necessary to empirically identify both intrinsically linear and nonlinear longitudinal CREMs. This study includes an applied example that utilizes the piecewise CREM with real data and three simulation studies to assess the data conditions necessary to estimate linear, quadratic, and piecewise CREMs. Results show that the number of repeated measurements collected on groups impacts the ability to recover the group effects. Additionally, functional form complexity impacted data collection requirements for estimating longitudinal CREMs.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 2","pages":"375-394"},"PeriodicalIF":1.8,"publicationDate":"2024-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12334","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139543995","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Statistical inference for agreement between multiple raters on a binary scale 二元量表上多个评分者之间一致性的统计推断。
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-01-17 DOI: 10.1111/bmsp.12333
Sophie Vanbelle

Agreement studies often involve more than two raters or repeated measurements. In the presence of two raters, the proportion of agreement and of positive agreement are simple and popular agreement measures for binary scales. These measures were generalized to agreement studies involving more than two raters with statistical inference procedures proposed on an empirical basis. We present two alternatives. The first is a Wald confidence interval using standard errors obtained by the delta method. The second involves Bayesian statistical inference not requiring any specific Bayesian software. These new procedures show better statistical behaviour than the confidence intervals initially proposed. In addition, we provide analytical formulas to determine the minimum number of persons needed for a given number of raters when planning an agreement study. All methods are implemented in the R package simpleagree and the Shiny app simpleagree.

一致性研究通常涉及两个以上的评分者或重复测量。在有两个评分者的情况下,二元量表的一致比例和积极一致比例是简单而常用的一致度量。根据经验提出的统计推论程序,这些测量方法被推广到涉及两个以上评分者的一致性研究中。我们提出了两种替代方案。第一种是使用德尔塔法获得的标准误差的沃尔德置信区间。第二种涉及贝叶斯统计推断,不需要任何特定的贝叶斯软件。与最初提出的置信区间相比,这些新程序显示出更好的统计性能。此外,我们还提供了分析公式,以便在计划协议研究时确定给定数量的评分者所需的最少人数。所有方法都在 R 软件包 simpleagree 和 Shiny 应用程序 simpleagree 中实现。
{"title":"Statistical inference for agreement between multiple raters on a binary scale","authors":"Sophie Vanbelle","doi":"10.1111/bmsp.12333","DOIUrl":"10.1111/bmsp.12333","url":null,"abstract":"<p>Agreement studies often involve more than two raters or repeated measurements. In the presence of two raters, the proportion of agreement and of positive agreement are simple and popular agreement measures for binary scales. These measures were generalized to agreement studies involving more than two raters with statistical inference procedures proposed on an empirical basis. We present two alternatives. The first is a Wald confidence interval using standard errors obtained by the delta method. The second involves Bayesian statistical inference not requiring any specific Bayesian software. These new procedures show better statistical behaviour than the confidence intervals initially proposed. In addition, we provide analytical formulas to determine the minimum number of persons needed for a given number of raters when planning an agreement study. All methods are implemented in the R package <i>simpleagree</i> and the Shiny app <i>simpleagree</i>.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 2","pages":"245-260"},"PeriodicalIF":1.8,"publicationDate":"2024-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12333","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139486878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A cluster differences unfolding method for large datasets of preference ratings on an interval scale: Minimizing the mean squared centred residuals 用于区间尺度偏好评分大型数据集的聚类差异展开法:最小化居中残差均方。
IF 1.8 3区 心理学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2024-01-11 DOI: 10.1111/bmsp.12332
Rodrigo Macías, J. Fernando Vera, Willem J. Heiser

Clustering and spatial representation methods are often used in combination, to analyse preference ratings when a large number of individuals and/or object is involved. When analysed under an unfolding model, row-conditional linear transformations are usually most appropriate when the goal is to determine clusters of individuals with similar preferences. However, a significant problem with transformations that include both slope and intercept is the occurrence of degenerate solutions. In this paper, we propose a least squares unfolding method that performs clustering of individuals while simultaneously estimating the location of cluster centres and object locations in low-dimensional space. The method is based on minimising the mean squared centred residuals of the preference ratings with respect to the distances between cluster centres and object locations. At the same time, the distances are row-conditionally transformed with optimally estimated slope parameters. It is computationally efficient for large datasets, and does not suffer from the appearance of degenerate solutions. The performance of the method is analysed in an extensive Monte Carlo experiment. It is illustrated for a real data set and the results are compared with those obtained using a two-step clustering and unfolding procedure.

当涉及大量个体和/或对象时,聚类和空间表示方法通常会结合使用,以分析偏好评级。在展开模型下进行分析时,当目标是确定具有相似偏好的个体聚类时,行条件线性变换通常是最合适的。然而,同时包含斜率和截距的变换的一个重要问题是会出现退化解。在本文中,我们提出了一种最小二乘展开法,在对个体进行聚类的同时,还能估计聚类中心的位置和低维空间中的对象位置。该方法基于最小化偏好评级与聚类中心和对象位置之间距离的均方中心残差。同时,利用最优估计的斜率参数对距离进行行条件变换。该方法对大型数据集的计算效率很高,而且不会出现退化解。通过大量的蒙特卡罗实验分析了该方法的性能。对一个真实数据集进行了说明,并将结果与使用两步聚类和展开程序获得的结果进行了比较。
{"title":"A cluster differences unfolding method for large datasets of preference ratings on an interval scale: Minimizing the mean squared centred residuals","authors":"Rodrigo Macías,&nbsp;J. Fernando Vera,&nbsp;Willem J. Heiser","doi":"10.1111/bmsp.12332","DOIUrl":"10.1111/bmsp.12332","url":null,"abstract":"<p>Clustering and spatial representation methods are often used in combination, to analyse preference ratings when a large number of individuals and/or object is involved. When analysed under an unfolding model, row-conditional linear transformations are usually most appropriate when the goal is to determine clusters of individuals with similar preferences. However, a significant problem with transformations that include both slope and intercept is the occurrence of degenerate solutions. In this paper, we propose a least squares unfolding method that performs clustering of individuals while simultaneously estimating the location of cluster centres and object locations in low-dimensional space. The method is based on minimising the mean squared centred residuals of the preference ratings with respect to the distances between cluster centres and object locations. At the same time, the distances are row-conditionally transformed with optimally estimated slope parameters. It is computationally efficient for large datasets, and does not suffer from the appearance of degenerate solutions. The performance of the method is analysed in an extensive Monte Carlo experiment. It is illustrated for a real data set and the results are compared with those obtained using a two-step clustering and unfolding procedure.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 2","pages":"356-374"},"PeriodicalIF":1.8,"publicationDate":"2024-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139426139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
British Journal of Mathematical & Statistical Psychology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1