首页 > 最新文献

Applied Psychological Measurement最新文献

英文 中文
Calibrating Multidimensional Assessments With Structural Missingness: An Application of a Multiple-Group Higher-Order IRT Model. 用结构缺失校正多维评估:多组高阶IRT模型的应用。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2026-01-07 DOI: 10.1177/01466216251415011
Yale Quan, Chun Wang

Educational Constructs are becoming increasingly complex and are often conceptualized at both a general level and a subdomain level. It is often desirable to report scores from both levels simultaneously. However, to measure such complex constructs, a very large item bank that is hard for a student to complete in any reasonable timeframe is needed. Furthermore, most current score reporting practices either only report subdomain scores, or the general domain score is calculated post hoc. We propose that a multiple group HO-IRT model with structural missingness can be used to simultaneously report general and subdomain scores while controlling assessment length. Although the model itself is not new, we consider a novel application scenario using a NEAT design with both a representative and non-representative anchor test. While a representative anchor test is recommended in literature, it is sometimes unrealistic in practice when the multidimensional construct shifts over time. Hence, exploring the parameter recovery of multiple group HO-IRT in the presence of non-representative anchor test is especially interesting and important. We show, through Monte Carlo simulation, that the RMSE of IRT estimates retrieved under a non-representative anchor item set with a moderate correlation between the higher- and lower-order factors, is comparable to the RMSE of IRT estimates retrieved under a representative anchor item set. Missing data were addressed using a full-information maximum likelihood approach to parameter estimation.

教育结构正变得越来越复杂,并且经常在一般级别和子领域级别概念化。同时报告两个级别的分数通常是可取的。然而,要测量如此复杂的结构,需要一个非常大的题库,学生很难在任何合理的时间框架内完成。此外,大多数当前的分数报告实践要么只报告子领域分数,要么在事后计算一般领域分数。我们提出了一个具有结构缺失的多组HO-IRT模型,可以在控制评估长度的同时报告一般和子域分数。虽然模型本身并不新鲜,但我们考虑了一个使用具有代表性和非代表性锚点测试的NEAT设计的新颖应用场景。虽然有代表性的锚点测试在文献中被推荐,但在实践中,当多维结构随时间变化时,它有时是不现实的。因此,探索多组HO-IRT在非代表性锚检验下的参数恢复就显得尤为有趣和重要。我们通过蒙特卡罗模拟表明,在非代表性锚项目集下检索的IRT估计的RMSE与在代表性锚项目集下检索的IRT估计的RMSE具有较高和较低阶因素之间的适度相关性,可与代表性锚项目集检索的IRT估计的RMSE相比较。使用全信息最大似然方法对缺失数据进行参数估计。
{"title":"Calibrating Multidimensional Assessments With Structural Missingness: An Application of a Multiple-Group Higher-Order IRT Model.","authors":"Yale Quan, Chun Wang","doi":"10.1177/01466216251415011","DOIUrl":"10.1177/01466216251415011","url":null,"abstract":"<p><p>Educational Constructs are becoming increasingly complex and are often conceptualized at both a general level and a subdomain level. It is often desirable to report scores from both levels simultaneously. However, to measure such complex constructs, a very large item bank that is hard for a student to complete in any reasonable timeframe is needed. Furthermore, most current score reporting practices either only report subdomain scores, or the general domain score is calculated post hoc. We propose that a multiple group HO-IRT model with structural missingness can be used to simultaneously report general and subdomain scores while controlling assessment length. Although the model itself is not new, we consider a novel application scenario using a NEAT design with both a representative and non-representative anchor test. While a representative anchor test is recommended in literature, it is sometimes unrealistic in practice when the multidimensional construct shifts over time. Hence, exploring the parameter recovery of multiple group HO-IRT in the presence of non-representative anchor test is especially interesting and important. We show, through Monte Carlo simulation, that the RMSE of IRT estimates retrieved under a non-representative anchor item set with a moderate correlation between the higher- and lower-order factors, is comparable to the RMSE of IRT estimates retrieved under a representative anchor item set. Missing data were addressed using a full-information maximum likelihood approach to parameter estimation.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251415011"},"PeriodicalIF":1.2,"publicationDate":"2026-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12779540/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145953603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Latent Trait Estimation in Multidimensional Forced Choice Measures: Latent Regression Multi-Unidimensional Pairwise Preference Model. 改进多维强迫选择度量中的潜在特征估计:潜在回归多单维两两偏好模型。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2026-01-03 DOI: 10.1177/01466216251415189
Sean Joo, Philseok Lee, Stephen Stark

The field of psychometrics has made remarkable progress in developing item response theory (IRT) models for analyzing multidimensional forced choice (MFC) measures. This study introduces an innovative method that enhances the latent trait estimation of the Multi-Unidimensional Pairwise Preference (MUPP) model by incorporating latent regression modeling. To validate the efficacy of the new method, we conducted a comprehensive simulation study. The results of the study provide compelling evidence that the proposed latent regression MUPP (LR-MUPP) model significantly improves the accuracy of the latent trait estimation. This study opens new avenues for future research and encourages further development and refinement of MFC IRT models and their applications.

心理测量学领域在建立项目反应理论(IRT)模型来分析多维强迫选择(MFC)测量方面取得了显著进展。本研究提出了一种创新的方法,结合潜在回归模型,提高了多维配对偏好(MUPP)模型的潜在性状估计。为了验证新方法的有效性,我们进行了全面的仿真研究。研究结果有力地证明,所提出的潜在回归MUPP (LR-MUPP)模型显著提高了潜在性状估计的准确性。本研究为未来的研究开辟了新的途径,并鼓励了MFC IRT模型及其应用的进一步发展和完善。
{"title":"Improving Latent Trait Estimation in Multidimensional Forced Choice Measures: Latent Regression Multi-Unidimensional Pairwise Preference Model.","authors":"Sean Joo, Philseok Lee, Stephen Stark","doi":"10.1177/01466216251415189","DOIUrl":"10.1177/01466216251415189","url":null,"abstract":"<p><p>The field of psychometrics has made remarkable progress in developing item response theory (IRT) models for analyzing multidimensional forced choice (MFC) measures. This study introduces an innovative method that enhances the latent trait estimation of the Multi-Unidimensional Pairwise Preference (MUPP) model by incorporating latent regression modeling. To validate the efficacy of the new method, we conducted a comprehensive simulation study. The results of the study provide compelling evidence that the proposed latent regression MUPP (LR-MUPP) model significantly improves the accuracy of the latent trait estimation. This study opens new avenues for future research and encourages further development and refinement of MFC IRT models and their applications.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251415189"},"PeriodicalIF":1.2,"publicationDate":"2026-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12764422/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145907257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Can the Generalized Graded Unfolding Model Fit Dominance Responses? 广义梯度展开模型能拟合优势反应吗?
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-12-07 DOI: 10.1177/01466216251401214
Jianbin Fu, Xuan Tan, Patrick C Kyllonen

Theoretically, the generalized graded unfolding model (GGUM) is more flexible than the generalized partial credit model (GPCM), a dominance model. For item responses generated by the GPCM, the GGUM estimations can generate overlapping item response curves with those from the GPCM over a range of latent trait scores covering almost all of the population. The discrimination and category threshold estimates from the two models are approximately equal. It is necessary to use an informative prior around an extreme location (e.g., 4 for a positive GPCM item) or fix the extreme locations in the GGUM estimation of GPCM items to achieve the desired estimation. The simulation study and the applications on two real datasets support the theoretical claims. Various practical implications are discussed, and suggestions for future research are provided.

从理论上讲,广义分级展开模型(GGUM)比优势模型广义部分信用模型(GPCM)更具灵活性。对于由GPCM生成的项目反应,GGUM估计可以在几乎覆盖所有群体的潜在特质得分范围内生成与GPCM的项目反应重叠的曲线。两种模型的判别和类别阈值估计大致相等。有必要在极端位置周围使用信息先验(例如,4为正GPCM项目)或固定GPCM项目的GGUM估计中的极端位置以实现所需的估计。仿真研究和在两个实际数据集上的应用支持了理论观点。讨论了各种实际意义,并对未来的研究提出了建议。
{"title":"Can the Generalized Graded Unfolding Model Fit Dominance Responses?","authors":"Jianbin Fu, Xuan Tan, Patrick C Kyllonen","doi":"10.1177/01466216251401214","DOIUrl":"10.1177/01466216251401214","url":null,"abstract":"<p><p>Theoretically, the generalized graded unfolding model (GGUM) is more flexible than the generalized partial credit model (GPCM), a dominance model. For item responses generated by the GPCM, the GGUM estimations can generate overlapping item response curves with those from the GPCM over a range of latent trait scores covering almost all of the population. The discrimination and category threshold estimates from the two models are approximately equal. It is necessary to use an informative prior around an extreme location (e.g., 4 for a positive GPCM item) or fix the extreme locations in the GGUM estimation of GPCM items to achieve the desired estimation. The simulation study and the applications on two real datasets support the theoretical claims. Various practical implications are discussed, and suggestions for future research are provided.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251401214"},"PeriodicalIF":1.2,"publicationDate":"2025-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12682685/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145716376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Unreliability of Test-Retest Reliability. 论重测信度的不信度。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-11-26 DOI: 10.1177/01466216251401213
Domenic Groh

The Test-Retest Coefficient (TRC) is a central metric of reliability in Classical Test Theory and modern psychological assessments. Originally developed by early 20th-century psychometricians, it relies on the assumptions of fixed (i.e., perfectly stable) true scores and independent error scores. However, these assumptions are rarely, if ever, tested, despite the fact that their violation can introduce significant biases. This article explores the foundations of these assumptions and examines the performance of the TRC under varying conditions, including different sample sizes, true score stability, and error score dependence. Using simulated data, results show that decreasing true score stability biases TRC estimates, leading to underestimations of reliability. Additionally, error score dependence can inflate TRC values, making unreliable measures appear reliable. More fundamentally, when these assumptions are violated, the TRC becomes underidentified, meaning that multiple, substantively different data-generating processes can yield the same coefficient, thus undermining its interpretability. These findings call into question the TRC's suitability for applied settings, especially when traits fluctuate over time or measurement conditions are uncontrolled. Alternative approaches are briefly discussed.

在经典测试理论和现代心理评估中,重测系数(TRC)是衡量信度的核心指标。它最初是由20世纪初的心理测量学家开发的,它依赖于固定(即完全稳定)的真实分数和独立误差分数的假设。然而,这些假设很少(如果有的话)得到检验,尽管违反这些假设可能会带来显著的偏差。本文探讨了这些假设的基础,并检查了TRC在不同条件下的性能,包括不同的样本量、真实分数稳定性和错误分数依赖性。使用模拟数据,结果表明,真实分数稳定性的降低会导致TRC估计偏差,从而导致可靠性的低估。此外,误差分数依赖性会使TRC值膨胀,使不可靠的度量看起来可靠。更根本的是,当这些假设被违反时,TRC就会被低估,这意味着多个实质上不同的数据生成过程可以产生相同的系数,从而破坏了其可解释性。这些发现对TRC对应用设置的适用性提出了质疑,特别是当性状随时间波动或测量条件不受控制时。简要讨论了各种备选方法。
{"title":"On the Unreliability of Test-Retest Reliability.","authors":"Domenic Groh","doi":"10.1177/01466216251401213","DOIUrl":"10.1177/01466216251401213","url":null,"abstract":"<p><p>The Test-Retest Coefficient (TRC) is a central metric of reliability in Classical Test Theory and modern psychological assessments. Originally developed by early 20th-century psychometricians, it relies on the assumptions of fixed (i.e., perfectly stable) true scores and independent error scores. However, these assumptions are rarely, if ever, tested, despite the fact that their violation can introduce significant biases. This article explores the foundations of these assumptions and examines the performance of the TRC under varying conditions, including different sample sizes, true score stability, and error score dependence. Using simulated data, results show that decreasing true score stability biases TRC estimates, leading to underestimations of reliability. Additionally, error score dependence can inflate TRC values, making unreliable measures appear reliable. More fundamentally, when these assumptions are violated, the TRC becomes underidentified, meaning that multiple, substantively different data-generating processes can yield the same coefficient, thus undermining its interpretability. These findings call into question the TRC's suitability for applied settings, especially when traits fluctuate over time or measurement conditions are uncontrolled. Alternative approaches are briefly discussed.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251401213"},"PeriodicalIF":1.2,"publicationDate":"2025-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12657207/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145649801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anchor Detection Strategy in Moderated Non-Linear Factor Analysis for Differential Item Functioning (DIF). 差异项目功能(DIF)的调节非线性因子分析中的锚点检测策略。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-11-24 DOI: 10.1177/01466216251401206
Sooyong Lee, Suyoung Kim, Seung W Choi

Ensuring measurement invariance is crucial for fair psychological and educational assessments, particularly in detecting Differential Item Functioning (DIF). Moderated Non-linear Factor Analysis (MNLFA) provides a flexible framework for detecting DIF by modeling item parameters as functions of observed covariates. However, a significant challenge in MNLFA-based DIF detection is anchor item selection, as improperly chosen anchors can bias results. This study proposes a refined constrained-baseline anchor detection approach utilizing information criteria (IC) for model selection. The proposed three-step procedure sequentially identifies potential DIF items through the Bayesian Information Criterion (BIC) and Weighted Information Criterion (WIC), followed by DIF-free anchor items using the Akaike Information Criterion (AIC). The method's effectiveness in controlling Type I error rates while maintaining statistical power is evaluated through simulation studies and empirical data analysis. Comparisons with regularization approaches demonstrate the proposed method's accuracy and computational efficiency.

确保测量的不变性对于公平的心理和教育评估至关重要,特别是在检测差异项目功能(DIF)方面。调节非线性因子分析(MNLFA)通过将项目参数建模为观测协变量的函数,为检测DIF提供了一个灵活的框架。然而,在基于mnlfa的DIF检测中,一个重要的挑战是锚点项目的选择,因为锚点选择不当会导致结果偏差。本研究提出了一种利用信息标准(IC)进行模型选择的改进约束基线锚点检测方法。该方法通过贝叶斯信息准则(BIC)和加权信息准则(WIC)确定潜在的DIF项目,然后使用赤池信息准则(AIC)确定无DIF的锚点项目。通过仿真研究和实证数据分析,评价了该方法在保持统计威力的同时控制I类错误率的有效性。与正则化方法的比较证明了该方法的准确性和计算效率。
{"title":"Anchor Detection Strategy in Moderated Non-Linear Factor Analysis for Differential Item Functioning (DIF).","authors":"Sooyong Lee, Suyoung Kim, Seung W Choi","doi":"10.1177/01466216251401206","DOIUrl":"https://doi.org/10.1177/01466216251401206","url":null,"abstract":"<p><p>Ensuring measurement invariance is crucial for fair psychological and educational assessments, particularly in detecting Differential Item Functioning (DIF). Moderated Non-linear Factor Analysis (MNLFA) provides a flexible framework for detecting DIF by modeling item parameters as functions of observed covariates. However, a significant challenge in MNLFA-based DIF detection is anchor item selection, as improperly chosen anchors can bias results. This study proposes a refined constrained-baseline anchor detection approach utilizing information criteria (IC) for model selection. The proposed three-step procedure sequentially identifies potential DIF items through the Bayesian Information Criterion (BIC) and Weighted Information Criterion (WIC), followed by DIF-free anchor items using the Akaike Information Criterion (AIC). The method's effectiveness in controlling Type I error rates while maintaining statistical power is evaluated through simulation studies and empirical data analysis. Comparisons with regularization approaches demonstrate the proposed method's accuracy and computational efficiency.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251401206"},"PeriodicalIF":1.2,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12643905/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145641338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distinguishing Between Models for Extreme and Midpoint Response Styles as Opposite Poles of a Single Dimension versus Two Separate Dimensions: A Simulation Study. 区分极端和中点响应风格模型作为单一维度与两个独立维度的对立极点:一项模拟研究。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-09-13 DOI: 10.1177/01466216251379471
Martijn Schoenmakers, Maria Bolsinova, Jesper Tijmstra

Extreme and midpoint response styles have frequently been found to decrease the validity of Likert-type questionnaire results. Different approaches for modelling extreme and midpoint responding have been proposed in the literature, with some advocating for a unidimensional conceptualization of the response styles as opposite poles, and others modelling them as separate dimensions. How these response styles are modelled influences the estimation complexity, parameter estimates, and detection of and correction for response styles in IRT models. For these reasons, we examine if it is possible to empirically distinguish between extreme and midpoint responding as two separate dimensions versus two opposite sides of a single dimension. The various conceptualizations are modelled using the multidimensional nominal response model, with the AIC and BIC being used to distinguish between the competing models in a simulation study and an empirical example. Results indicate good performance of both information criteria given sufficient sample size, test length, and response style strength. The BIC outperformed the AIC in cases where no response styles were present, while the AIC outperformed the BIC in cases where multiple response style dimensions were present. Implications of the results for practice are discussed.

极端和中点反应风格经常被发现会降低李克特型问卷结果的效度。文献中提出了不同的模拟极端和中点反应的方法,一些人主张将反应风格作为相反的两极进行一维概念化,而另一些人则将它们作为单独的维度进行建模。如何对这些响应样式建模会影响IRT模型中响应样式的估计复杂性、参数估计以及检测和校正。由于这些原因,我们检查是否有可能在经验上区分极端和中点响应作为两个独立的维度与单个维度的两个相反的方面。使用多维名义响应模型对各种概念化进行建模,AIC和BIC用于区分模拟研究和经验示例中的竞争模型。结果表明,在给定足够的样本量、测试长度和响应风格强度的情况下,这两种信息标准都具有良好的性能。在没有反应风格维度的情况下,BIC优于AIC,而在存在多个反应风格维度的情况下,AIC优于BIC。讨论了研究结果对实践的影响。
{"title":"Distinguishing Between Models for Extreme and Midpoint Response Styles as Opposite Poles of a Single Dimension versus Two Separate Dimensions: A Simulation Study.","authors":"Martijn Schoenmakers, Maria Bolsinova, Jesper Tijmstra","doi":"10.1177/01466216251379471","DOIUrl":"10.1177/01466216251379471","url":null,"abstract":"<p><p>Extreme and midpoint response styles have frequently been found to decrease the validity of Likert-type questionnaire results. Different approaches for modelling extreme and midpoint responding have been proposed in the literature, with some advocating for a unidimensional conceptualization of the response styles as opposite poles, and others modelling them as separate dimensions. How these response styles are modelled influences the estimation complexity, parameter estimates, and detection of and correction for response styles in IRT models. For these reasons, we examine if it is possible to empirically distinguish between extreme and midpoint responding as two separate dimensions versus two opposite sides of a single dimension. The various conceptualizations are modelled using the multidimensional nominal response model, with the AIC and BIC being used to distinguish between the competing models in a simulation study and an empirical example. Results indicate good performance of both information criteria given sufficient sample size, test length, and response style strength. The BIC outperformed the AIC in cases where no response styles were present, while the AIC outperformed the BIC in cases where multiple response style dimensions were present. Implications of the results for practice are discussed.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251379471"},"PeriodicalIF":1.2,"publicationDate":"2025-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12433433/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145070752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
fcirt: An R Package for Forced Choice Models in Item Response Theory. 第一章:项目反应理论中强迫选择模型的R包。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-09-10 DOI: 10.1177/01466216251378771
Naidan Tu, Sean Joo, Philseok Lee, Stephen Stark

Multidimensional forced choice (MFC) formats have emerged as a promising alternative to traditional single statement Likert-type measures for assessing noncognitive traits while reducing response biases. As MFC formats become more widely used, there is a growing need for tools to support MFC analysis, which motivated the development of the fcirt package. The fcirt package estimates forced choice model parameters using Bayesian methods. It currently enables estimation of the Generalized Graded Unfolding Model (GGUM; Roberts et al., 2000)-based Multi-Unidimensional Pairwise Preference (MUPP) model using rstan, which implements the Hamiltonian Monte Carlo (HMC) sampling algorithm. fcirt also includes functions for computing item and test information functions to evaluate the quality of MFC assessments, as well as functions for Bayesian diagnostic plotting to assist with model evaluation and convergence assessment.

多维强迫选择(MFC)格式已经成为传统的单语句李克特测量方法的一个有希望的替代方案,用于评估非认知特征,同时减少反应偏差。随着MFC格式的广泛使用,越来越需要支持MFC分析的工具,这推动了第一个包的开发。第一个包使用贝叶斯方法估计强制选择模型参数。目前,它可以使用rstan估计基于广义梯度展开模型(GGUM; Roberts et al., 2000)的多维配对偏好(MUPP)模型,该模型实现了哈密顿蒙特卡罗(HMC)采样算法。fcirt还包括计算项目和测试信息函数的功能,以评估MFC评估的质量,以及贝叶斯诊断绘图的功能,以协助模型评估和收敛性评估。
{"title":"<i>fcirt</i>: An R Package for Forced Choice Models in Item Response Theory.","authors":"Naidan Tu, Sean Joo, Philseok Lee, Stephen Stark","doi":"10.1177/01466216251378771","DOIUrl":"10.1177/01466216251378771","url":null,"abstract":"<p><p>Multidimensional forced choice (MFC) formats have emerged as a promising alternative to traditional single statement Likert-type measures for assessing noncognitive traits while reducing response biases. As MFC formats become more widely used, there is a growing need for tools to support MFC analysis, which motivated the development of the <i>fcirt</i> package. The <i>fcirt</i> package estimates forced choice model parameters using Bayesian methods. It currently enables estimation of the Generalized Graded Unfolding Model (GGUM; Roberts et al., 2000)-based Multi-Unidimensional Pairwise Preference (MUPP) model using <i>rstan</i>, which implements the Hamiltonian Monte Carlo (HMC) sampling algorithm. <i>fcirt</i> also includes functions for computing item and test information functions to evaluate the quality of MFC assessments, as well as functions for Bayesian diagnostic plotting to assist with model evaluation and convergence assessment.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251378771"},"PeriodicalIF":1.2,"publicationDate":"2025-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12423082/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145065903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Generation of Rule-Based Raven-Like Matrices in R: The matRiks Package. 在R中自动生成基于规则的类乌鸦矩阵:矩阵包。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-09-02 DOI: 10.1177/01466216251374826
Andrea Brancaccio, Ottavia M Epifania, Pasquale Anselmi, Debora de Chiusole
{"title":"Automatic Generation of Rule-Based Raven-Like Matrices in R: The matRiks Package.","authors":"Andrea Brancaccio, Ottavia M Epifania, Pasquale Anselmi, Debora de Chiusole","doi":"10.1177/01466216251374826","DOIUrl":"10.1177/01466216251374826","url":null,"abstract":"","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251374826"},"PeriodicalIF":1.2,"publicationDate":"2025-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12405199/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145001708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CALMs: A Shiny Application for Comprehensive Analysis of Latent Means. CALMs:潜在均值综合分析的闪亮应用。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-08-31 DOI: 10.1177/01466216251371173
Kim Nimon, Julia Fulmore, Gregg Keiffer, Bryn Hammack-Brown

This article presents a Shiny application CALMs for comprehensively comparing groups via latent means, which includes the examination of group equivalency, propensity score analysis, measurement invariance analysis, and assessment of latent mean differences of equivalent groups with invariant data. Despite the importance of these techniques, their application can be complex and time-consuming, particularly for researchers not experienced in advanced statistical methods. The Shiny application CALMs makes this cumbersome process more accessible to a broader range of users. In addition, it allows researchers to focus more on the interpretation aspect of the research rather than the labor required for testing. The practical utility of the CALMs application is demonstrated using real-world data, highlighting the potential of the application to enhance the validity and reliability of group comparison studies in psychological research.

本文介绍了一种利用潜在均值综合比较群体的新应用CALMs,包括群体等效性检验、倾向得分分析、测量不变性分析以及用不变数据评估等效群体的潜在均值差异。尽管这些技术很重要,但它们的应用可能是复杂和耗时的,特别是对于没有高级统计方法经验的研究人员。Shiny的应用程序CALMs使这个繁琐的过程对更广泛的用户更容易访问。此外,它允许研究人员更多地关注研究的解释方面,而不是测试所需的劳动力。使用真实世界的数据演示了CALMs应用程序的实际效用,突出了该应用程序在提高心理学研究中群体比较研究的有效性和可靠性方面的潜力。
{"title":"CALMs: A Shiny Application for Comprehensive Analysis of Latent Means.","authors":"Kim Nimon, Julia Fulmore, Gregg Keiffer, Bryn Hammack-Brown","doi":"10.1177/01466216251371173","DOIUrl":"https://doi.org/10.1177/01466216251371173","url":null,"abstract":"<p><p>This article presents a Shiny application CALMs for comprehensively comparing groups via latent means, which includes the examination of group equivalency, propensity score analysis, measurement invariance analysis, and assessment of latent mean differences of equivalent groups with invariant data. Despite the importance of these techniques, their application can be complex and time-consuming, particularly for researchers not experienced in advanced statistical methods. The Shiny application CALMs makes this cumbersome process more accessible to a broader range of users. In addition, it allows researchers to focus more on the interpretation aspect of the research rather than the labor required for testing. The practical utility of the CALMs application is demonstrated using real-world data, highlighting the potential of the application to enhance the validity and reliability of group comparison studies in psychological research.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251371173"},"PeriodicalIF":1.2,"publicationDate":"2025-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12399567/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144993935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
implicitMeasures: An R Package for Scoring the Implicit Association Test and the Single-Category Implicit Association Test. 内隐关联测验和单类别内隐关联测验的R量表。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-08-25 DOI: 10.1177/01466216251371532
Ottavia M Epifania, Pasquale Anselmi, Egidio Robusto
{"title":"implicitMeasures: An R Package for Scoring the Implicit Association Test and the Single-Category Implicit Association Test.","authors":"Ottavia M Epifania, Pasquale Anselmi, Egidio Robusto","doi":"10.1177/01466216251371532","DOIUrl":"https://doi.org/10.1177/01466216251371532","url":null,"abstract":"","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251371532"},"PeriodicalIF":1.2,"publicationDate":"2025-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12378261/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144974606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied Psychological Measurement
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1