Pablo Nájera, Rodrigo S. Kreitchmann, Scarlett Escudero, Francisco J. Abad, Jimmy de la Torre, Miguel A. Sorrel
Diagnostic classification modelling (DCM) is a family of restricted latent class models often used in educational settings to assess students' strengths and weaknesses. Recently, there has been growing interest in applying DCM to noncognitive traits in fields such as clinical and organizational psychology, as well as personality profiling. To address common response biases in these assessments, such as social desirability, Huang (2023, Educational and Psychological Measurement, 83, 146) adopted the forced-choice (FC) item format within the DCM framework, developing the FC-DCM. This model assumes that examinees with no clear preference for any statements in an FC block will choose completely at random. Additionally, the unique parametrization of the FC-DCM poses challenges for integration with established DCM frameworks in the literature. In the present study, we enhance the capabilities of DCM by introducing a general diagnostic framework for FC assessments. We present an adaptation of the G-DINA model to accommodate FC responses. Simulation results show that the G-DINA model provides accurate classifications, item parameter estimates and attribute correlations, outperforming the FC-DCM in realistic scenarios where item discrimination varies. A real FC assessment example further illustrates the better model fit of the G-DINA. Practical recommendations for using the FC format in diagnostic assessments of noncognitive traits are provided.
诊断分类模型(DCM)是一类受限的潜在类别模型,通常用于教育环境中评估学生的优势和劣势。最近,人们对将DCM应用于临床和组织心理学以及人格分析等领域的非认知特征越来越感兴趣。为了解决这些评估中常见的反应偏差,例如社会期望,Huang (2023, Educational and Psychological Measurement, 83,146)在DCM框架中采用了强制选择(FC)项目格式,开发了FC-DCM。该模型假设考生对FC块中的任何语句没有明确的偏好,将完全随机选择。此外,FC-DCM的独特参数化对与文献中已建立的DCM框架的集成提出了挑战。在本研究中,我们通过引入FC评估的一般诊断框架来增强DCM的能力。我们提出了一种适应G-DINA模型以适应FC响应。仿真结果表明,G-DINA模型提供了准确的分类、项目参数估计和属性相关性,在项目区分变化的现实场景中优于FC-DCM模型。一个实际的FC评估实例进一步说明了G-DINA模型拟合效果较好。提供了在非认知特征的诊断评估中使用FC格式的实用建议。
{"title":"A general diagnostic modelling framework for forced-choice assessments","authors":"Pablo Nájera, Rodrigo S. Kreitchmann, Scarlett Escudero, Francisco J. Abad, Jimmy de la Torre, Miguel A. Sorrel","doi":"10.1111/bmsp.12393","DOIUrl":"10.1111/bmsp.12393","url":null,"abstract":"<p>Diagnostic classification modelling (DCM) is a family of restricted latent class models often used in educational settings to assess students' strengths and weaknesses. Recently, there has been growing interest in applying DCM to noncognitive traits in fields such as clinical and organizational psychology, as well as personality profiling. To address common response biases in these assessments, such as social desirability, Huang (2023, <i>Educational and Psychological Measurement</i>, <i>83</i>, 146) adopted the forced-choice (FC) item format within the DCM framework, developing the FC-DCM. This model assumes that examinees with no clear preference for any statements in an FC block will choose completely at random. Additionally, the unique parametrization of the FC-DCM poses challenges for integration with established DCM frameworks in the literature. In the present study, we enhance the capabilities of DCM by introducing a general diagnostic framework for FC assessments. We present an adaptation of the G-DINA model to accommodate FC responses. Simulation results show that the G-DINA model provides accurate classifications, item parameter estimates and attribute correlations, outperforming the FC-DCM in realistic scenarios where item discrimination varies. A real FC assessment example further illustrates the better model fit of the G-DINA. Practical recommendations for using the FC format in diagnostic assessments of noncognitive traits are provided.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"78 3","pages":"1025-1047"},"PeriodicalIF":1.8,"publicationDate":"2025-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://bpspsychub.onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12393","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144035912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We present a novel approach for computing model scores for ordinal factor models, that is, graded response models (GRMs) fitted with a limited information (LI) estimator. The method makes it possible to compute score-based tests for parameter instability for ordinal factor models. This way, rapid execution of numerous parameter instability tests for multidimensional item response theory (MIRT) models is facilitated. We present a comparative analysis of the performance of the proposed score-based tests for ordinal factor models in comparison to tests for GRMs fitted with a full information (FI) estimator. The new method has a good Type I error rate, high power and is computationally faster than FI estimation. We further illustrate that the proposed method works well with complex models in real data applications. The method is implemented in the lavaan package in R.
{"title":"Score-based tests for parameter instability in ordinal factor models","authors":"Franz Classe, Rudolf Debelak, Christoph Kern","doi":"10.1111/bmsp.12392","DOIUrl":"10.1111/bmsp.12392","url":null,"abstract":"<p>We present a novel approach for computing model scores for ordinal factor models, that is, graded response models (GRMs) fitted with a limited information (LI) estimator. The method makes it possible to compute score-based tests for parameter instability for ordinal factor models. This way, rapid execution of numerous parameter instability tests for multidimensional item response theory (MIRT) models is facilitated. We present a comparative analysis of the performance of the proposed score-based tests for ordinal factor models in comparison to tests for GRMs fitted with a full information (FI) estimator. The new method has a good Type I error rate, high power and is computationally faster than FI estimation. We further illustrate that the proposed method works well with complex models in real data applications. The method is implemented in the <i>lavaan</i> package in R.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"78 3","pages":"996-1024"},"PeriodicalIF":1.8,"publicationDate":"2025-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://bpspsychub.onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12392","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144052855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dexin Shi, Bo Zhang, Wolfgang Wiedermann, Amanda J. Fairchild
Distinguishing cause from effect – that is, determining whether x causes y (x → y) or, alternatively, whether y causes x (y → x) – is a primary research goal in many psychological research areas. Despite its importance, determining causal direction with observational data remains a difficult task. In this study, we introduce an independence-based approach for causal discovery between two variables of interest under a linear non-Gaussian model framework. We propose a two-step algorithm based on distance correlations that provides empirical conclusions on the causal directionality of effects under realistic conditions typically seen in psychological studies, that is, in the presence of hidden confounders. The performance of the proposed algorithm is evaluated using Monte-Carlo simulations. Findings suggest that the algorithm can effectively detect the causal direction between two variables of interest, even in the presence of weak hidden confounders. Moreover, distance correlations provide useful insights into the magnitude of hidden confounding. We provide an empirical example to demonstrate the application of our proposed approach and discuss practical implications and future directions.
{"title":"Distinguishing cause from effect in psychological research: An independence-based approach under linear non-Gaussian models","authors":"Dexin Shi, Bo Zhang, Wolfgang Wiedermann, Amanda J. Fairchild","doi":"10.1111/bmsp.12391","DOIUrl":"10.1111/bmsp.12391","url":null,"abstract":"<p>Distinguishing cause from effect – that is, determining whether <i>x</i> causes <i>y</i> (<i>x</i> → <i>y</i>) or, alternatively, whether <i>y</i> causes <i>x</i> (<i>y</i> → <i>x</i>) – is a primary research goal in many psychological research areas. Despite its importance, determining causal direction with observational data remains a difficult task. In this study, we introduce an independence-based approach for causal discovery between two variables of interest under a linear non-Gaussian model framework. We propose a two-step algorithm based on distance correlations that provides empirical conclusions on the <i>causal directionality</i> of effects under realistic conditions typically seen in psychological studies, that is, in the presence of hidden confounders. The performance of the proposed algorithm is evaluated using Monte-Carlo simulations. Findings suggest that the algorithm can effectively detect the causal direction between two variables of interest, even in the presence of weak hidden confounders. Moreover, distance correlations provide useful insights into the magnitude of hidden confounding. We provide an empirical example to demonstrate the application of our proposed approach and discuss practical implications and future directions.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"78 3","pages":"965-995"},"PeriodicalIF":1.8,"publicationDate":"2025-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://bpspsychub.onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12391","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143998166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Good scientific practice requires that the reporting of the statistical analysis of experiments should include estimates of effect size as well as the results of tests of statistical significance. Good statistical practice requires that effect size estimates be reported along with some indication of their statistical uncertainty, such as a standard error. This article provides a review of effect sizes for experimental research, including expressions for the standard error of each effect size. It focuses on effect sizes for experiments with treatments having a single degree of freedom but also includes effect sizes for treatments with multiple degrees of freedom having either fixed or random effects.
{"title":"Effect sizes for experimental research","authors":"Larry V. Hedges","doi":"10.1111/bmsp.12389","DOIUrl":"10.1111/bmsp.12389","url":null,"abstract":"<p>Good scientific practice requires that the reporting of the statistical analysis of experiments should include estimates of effect size as well as the results of tests of statistical significance. Good statistical practice requires that effect size estimates be reported along with some indication of their statistical uncertainty, such as a standard error. This article provides a review of effect sizes for experimental research, including expressions for the standard error of each effect size. It focuses on effect sizes for experiments with treatments having a single degree of freedom but also includes effect sizes for treatments with multiple degrees of freedom having either fixed or random effects.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"79 1","pages":"31-45"},"PeriodicalIF":1.8,"publicationDate":"2025-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://bpspsychub.onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12389","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143755694","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Item preknowledge (IP) is a prevalent form of test fraud in educational assessment that can compromise test validity. Two common methods for detecting examinees with IP are score-differencing statistics and response similarity index (RSI). These statistics have different applications and respective advantages. In this paper, we propose a new method (Joint Survival Function Method,