首页 > 最新文献

Applied Psychological Measurement最新文献

英文 中文
Structure-Based Classification Approach. 基于结构的分类方法。
IF 1 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-14 DOI: 10.1177/01466216251360544
Jongwan Kim

This study introduces a novel structure-based classification (SBC) framework that leverages pairwise distance representations of rating data to enhance classification performance while mitigating individual differences in scale usage. Unlike conventional feature-based approaches that rely on absolute rating scores, SBC transforms rating data into structured representations by computing pairwise distances between rating dimensions. This transformation captures the relational structure of ratings, ensuring consistency between training and test datasets and enhancing model robustness. To evaluate the effectiveness of this approach, we conducted a simulation study in which participants rated stimuli across multiple affective dimensions, with systematic individual differences in scale usage. The results demonstrated that SBC successfully classified affective stimuli despite these variations, performing comparably to traditional classification methods. The findings suggest that relational structures among rating dimensions contain meaningful information for affective classification, akin to functional connectivity approaches in cognitive neuroscience. By focusing on rating interdependencies as well as absolute values, SBC provides a robust and generalizable method for analyzing subjective responses, with implications for psychological research.

本研究引入了一种新的基于结构的分类(SBC)框架,该框架利用评级数据的两两距离表示来提高分类性能,同时减轻量表使用的个体差异。与依赖绝对评分的传统基于特征的方法不同,SBC通过计算评分维度之间的成对距离将评分数据转换为结构化表示。这种转换捕获了评级的关系结构,确保了训练和测试数据集之间的一致性,增强了模型的鲁棒性。为了评估这种方法的有效性,我们进行了一项模拟研究,在该研究中,参与者在多个情感维度上对刺激进行评分,并在量表使用上存在系统的个体差异。结果表明,尽管存在这些差异,SBC仍能成功地对情感刺激进行分类,其表现与传统分类方法相当。研究结果表明,评级维度之间的关系结构包含情感分类的有意义信息,类似于认知神经科学中的功能连接方法。通过专注于评估相互依赖性和绝对值,SBC为分析主观反应提供了一种强大且可推广的方法,对心理学研究具有重要意义。
{"title":"Structure-Based Classification Approach.","authors":"Jongwan Kim","doi":"10.1177/01466216251360544","DOIUrl":"10.1177/01466216251360544","url":null,"abstract":"<p><p>This study introduces a novel structure-based classification (SBC) framework that leverages pairwise distance representations of rating data to enhance classification performance while mitigating individual differences in scale usage. Unlike conventional feature-based approaches that rely on absolute rating scores, SBC transforms rating data into structured representations by computing pairwise distances between rating dimensions. This transformation captures the relational structure of ratings, ensuring consistency between training and test datasets and enhancing model robustness. To evaluate the effectiveness of this approach, we conducted a simulation study in which participants rated stimuli across multiple affective dimensions, with systematic individual differences in scale usage. The results demonstrated that SBC successfully classified affective stimuli despite these variations, performing comparably to traditional classification methods. The findings suggest that relational structures among rating dimensions contain meaningful information for affective classification, akin to functional connectivity approaches in cognitive neuroscience. By focusing on rating interdependencies as well as absolute values, SBC provides a robust and generalizable method for analyzing subjective responses, with implications for psychological research.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251360544"},"PeriodicalIF":1.0,"publicationDate":"2025-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12264251/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144660749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Including Empirical Prior Information in the Reliable Change Index. 在可靠变化指数中加入经验先验信息。
IF 1 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-10 DOI: 10.1177/01466216251358492
R Philip Chalmers, Sarah Campbell

The reliable change index (RCI; Jacobson & Truax, 1991) is commonly used to assess whether individuals have changed across two measurement occasions, and has seen many augmentations and improvements since its initial conception. In this study, we extend an item response theory version of the RCI presented by Jabrayilov et al. (2016) by including empirical priors in the associated RCI computations whenever group-level differences are quantifiable given post-test response information. Based on a reanalysis and extension of a previous simulation study, we demonstrate that although a small amount of bias is added to the estimates of the latent trait differences when no true change is present, including empirical prior information will generally improve the Type I behavior of the model-based RCI. Consequently, when non-zero changes in the latent trait are present the bias and sampling variability are show to be more favorable than competing estimators, subsequently leading to an increase in power to detect non-zero changes.

可靠变化指数(RCI;Jacobson & Truax, 1991)通常用于评估个体是否在两个测量场合中发生了变化,并且自最初的概念以来已经看到了许多增强和改进。在本研究中,我们扩展了Jabrayilov等人(2016)提出的RCI的项目反应理论版本,在给定测试后反应信息的情况下,无论群体水平差异是可量化的,我们都会在相关的RCI计算中纳入经验先验。基于先前模拟研究的再分析和扩展,我们证明,尽管在没有真实变化的情况下,潜在特征差异的估计中添加了少量偏差,但包括经验先验信息通常会改善基于模型的RCI的I型行为。因此,当潜在特征的非零变化存在时,偏差和抽样可变性比竞争估计器更有利,随后导致检测非零变化的能力增加。
{"title":"Including Empirical Prior Information in the Reliable Change Index.","authors":"R Philip Chalmers, Sarah Campbell","doi":"10.1177/01466216251358492","DOIUrl":"10.1177/01466216251358492","url":null,"abstract":"<p><p>The reliable change index (RCI; Jacobson & Truax, 1991) is commonly used to assess whether individuals have changed across two measurement occasions, and has seen many augmentations and improvements since its initial conception. In this study, we extend an item response theory version of the RCI presented by Jabrayilov et al. (2016) by including empirical priors in the associated RCI computations whenever group-level differences are quantifiable given post-test response information. Based on a reanalysis and extension of a previous simulation study, we demonstrate that although a small amount of bias is added to the estimates of the latent trait differences when no true change is present, including empirical prior information will generally improve the Type I behavior of the model-based RCI. Consequently, when non-zero changes in the latent trait are present the bias and sampling variability are show to be more favorable than competing estimators, subsequently leading to an increase in power to detect non-zero changes.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251358492"},"PeriodicalIF":1.0,"publicationDate":"2025-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12245826/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144627476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using Group Differences in True Score Relationships to Evaluate Measurement Bias. 用真分关系的组差异评价测量偏倚。
IF 1 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-07 DOI: 10.1177/01466216251358491
Michael T Kane, Joanne Kane

This paper makes three contributions to our understanding of measurement bias and predictive bias in testing. First, we develop a linear model for assessing measurement bias across two tests and two groups in terms of the estimated true-score relationships between the two tests in the two groups. This new model for measurement bias is structurally similar to the Cleary model for predictive bias, but it relies on the Errors-in-Variables (EIV) regression model, rather than the Ordinary-Least-Squares (OLS) regression model. Second, we examine some differences between measurement bias and predictive bias in three cases in which two groups have different true-score means, and we illustrate how regression toward the mean in OLS regression can lead to questionable conclusions about test bias if the differences between measurement bias and predictive bias are ignored. Third, we reevaluate a body of empirical findings suggesting that the tests employed in college-admissions and employment-testing programs tend to over-predict criterion performance for minorities, and we show that these findings are consistent with the occurrence of substantial measurement bias against the minority group relative to the majority group.

本文对我们对测试中的测量偏差和预测偏差的理解做出了三个贡献。首先,我们建立了一个线性模型,根据两组中两个测试之间的估计真值关系来评估两个测试和两组之间的测量偏差。这种新的测量偏差模型在结构上类似于预测偏差的Cleary模型,但它依赖于变量误差(EIV)回归模型,而不是普通最小二乘(OLS)回归模型。其次,我们在两组真实得分均值不同的三种情况下检验了测量偏倚和预测偏倚之间的差异,并说明了如果忽略测量偏倚和预测偏倚之间的差异,OLS回归中的均值回归如何导致关于检验偏倚的可疑结论。第三,我们重新评估了一系列实证研究结果,这些研究结果表明,在大学录取和就业测试项目中使用的测试往往会高估少数族裔的标准表现,我们表明,这些发现与相对于多数群体而言,对少数族裔群体存在实质性的测量偏差是一致的。
{"title":"Using Group Differences in True Score Relationships to Evaluate Measurement Bias.","authors":"Michael T Kane, Joanne Kane","doi":"10.1177/01466216251358491","DOIUrl":"10.1177/01466216251358491","url":null,"abstract":"<p><p>This paper makes three contributions to our understanding of measurement bias and predictive bias in testing. First, we develop a linear model for assessing measurement bias across two tests and two groups in terms of the estimated true-score relationships between the two tests in the two groups. This new model for measurement bias is structurally similar to the Cleary model for predictive bias, but it relies on the Errors-in-Variables (EIV) regression model, rather than the Ordinary-Least-Squares (OLS) regression model. Second, we examine some differences between measurement bias and predictive bias in three cases in which two groups have different true-score means, and we illustrate how regression toward the mean in OLS regression can lead to questionable conclusions about test bias if the differences between measurement bias and predictive bias are ignored. Third, we reevaluate a body of empirical findings suggesting that the tests employed in college-admissions and employment-testing programs tend to over-predict criterion performance for minorities, and we show that these findings are consistent with the occurrence of substantial measurement bias against the minority group relative to the majority group.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251358491"},"PeriodicalIF":1.0,"publicationDate":"2025-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12234520/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144601949","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Standard Error Estimation for Subpopulation Non-invariance. 亚总体非不变性的标准误差估计。
IF 1 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-05 DOI: 10.1177/01466216251351947
Paul A Jewsbury

Score linking is widely used to place scores from different assessments, or the same assessment under different conditions, onto a common scale. A central concern is whether the linking function is invariant across subpopulations, as violations may threaten fairness. However, evaluating subpopulation differences in linked scores is challenging because linking error is not independent of sampling and measurement error when the same data are used to estimate the linking function and to compare score distributions. We show that common approaches involving neglecting linking error or treating it as independent substantially overestimate the standard errors of subpopulation differences. We introduce new methods that account for linking error dependencies. Simulation results demonstrate the accuracy of the proposed methods, and a practical example with real data illustrates how improved standard error estimation enhances power for detecting subpopulation non-invariance.

分数挂钩被广泛用于将不同评估的分数,或在不同条件下的同一评估的分数放在一个共同的尺度上。一个中心问题是,连接函数在各个子种群之间是否不变,因为违反连接函数可能会威胁到公平性。然而,当使用相同的数据来估计连接函数和比较分数分布时,评估关联分数的亚群体差异是具有挑战性的,因为连接误差并不独立于抽样和测量误差。我们表明,包括忽略连接误差或将其视为独立的常见方法实质上高估了亚群体差异的标准误差。我们引入了新的方法来解释错误依赖关系的链接。仿真结果表明了所提方法的准确性,并用一个实际数据实例说明了改进的标准误差估计提高了检测亚种群非不变性的能力。
{"title":"Standard Error Estimation for Subpopulation Non-invariance.","authors":"Paul A Jewsbury","doi":"10.1177/01466216251351947","DOIUrl":"10.1177/01466216251351947","url":null,"abstract":"<p><p>Score linking is widely used to place scores from different assessments, or the same assessment under different conditions, onto a common scale. A central concern is whether the linking function is invariant across subpopulations, as violations may threaten fairness. However, evaluating subpopulation differences in linked scores is challenging because linking error is not independent of sampling and measurement error when the same data are used to estimate the linking function and to compare score distributions. We show that common approaches involving neglecting linking error or treating it as independent substantially overestimate the standard errors of subpopulation differences. We introduce new methods that account for linking error dependencies. Simulation results demonstrate the accuracy of the proposed methods, and a practical example with real data illustrates how improved standard error estimation enhances power for detecting subpopulation non-invariance.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251351947"},"PeriodicalIF":1.0,"publicationDate":"2025-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12228644/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144585323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Experimental Design to Investigate Item Parameter Drift. 研究项目参数漂移的实验设计。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-01 Epub Date: 2025-01-24 DOI: 10.1177/01466216251316282
Peter Baldwin, Irina Grabovsky, Kimberly A Swygert, Thomas Fogle, Pilar Reid, Brian E Clauser

Methods for detecting item parameter drift may be inadequate when every exposed item is at risk for drift. To address this scenario, a strategy for detecting item parameter drift is proposed that uses only unexposed items deployed in a stratified random method within an experimental design. The proposed method is illustrated by investigating unexpected score increases on a high-stakes licensure exam. Results for this example were suggestive of item parameter drift but not significant at the .05 level.

当每个暴露的项目都有漂移的危险时,检测项目参数漂移的方法可能是不够的。为了解决这种情况,提出了一种检测项目参数漂移的策略,该策略仅使用实验设计中分层随机方法部署的未暴露项目。通过调查高风险执照考试中意外的分数增长来说明所提出的方法。本例的结果提示项目参数漂移,但在0.05水平上不显著。
{"title":"An Experimental Design to Investigate Item Parameter Drift.","authors":"Peter Baldwin, Irina Grabovsky, Kimberly A Swygert, Thomas Fogle, Pilar Reid, Brian E Clauser","doi":"10.1177/01466216251316282","DOIUrl":"10.1177/01466216251316282","url":null,"abstract":"<p><p>Methods for detecting item parameter drift may be inadequate when every exposed item is at risk for drift. To address this scenario, a strategy for detecting item parameter drift is proposed that uses only unexposed items deployed in a stratified random method within an experimental design. The proposed method is illustrated by investigating unexpected score increases on a high-stakes licensure exam. Results for this example were suggestive of item parameter drift but not significant at the .05 level.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"212-223"},"PeriodicalIF":1.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11760077/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143048347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detecting DIF with the Multi-Unidimensional Pairwise Preference Model: Lord's Chi-square and IPR-NCDIF Methods. 用多维配对偏好模型检测DIF: Lord卡方和IPR-NCDIF方法。
IF 1 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-01 DOI: 10.1177/01466216251351949
Lavanya S Kumar, Naidan Tu, Sean Joo, Stephen Stark

Multidimensional forced choice (MFC) measures are gaining prominence in noncognitive assessment. Yet there has been little research on detecting differential item functioning (DIF) with models for forced choice measures. This research extended two well-known DIF detection methods to MFC measures. Specifically, the performance of Lord's chi-square and item parameter replication (IPR) methods with MFC tests based on the Multi-Unidimensional Pairwise Preference (MUPP) model was investigated. The Type I error rate and power of the DIF detection methods were examined in a Monte Carlo simulation that manipulated sample size, impact, DIF source, and DIF magnitude. Both methods showed consistent power and were found to control Type I error well across study conditions, indicating that established approaches to DIF detection work well with the MUPP model. Lord's chi-square outperformed the IPR method when DIF source was statement discrimination while the opposite was true when DIF source was statement threshold. Also, both methods performed similarly and showed better power when DIF source was statement location, in line with previous research. Study implications and practical recommendations for DIF detection with MFC tests, as well as limitations, are discussed.

多维强迫选择(MFC)方法在非认知评估中越来越受到重视。然而,利用强迫选择测量模型检测差异项目功能(DIF)的研究很少。本研究将两种著名的DIF检测方法扩展到MFC测量中。具体而言,研究了基于多维成对偏好(MUPP)模型的MFC检验的Lord’s卡方和项目参数复制(IPR)方法的性能。在蒙特卡罗模拟中检验了I型错误率和DIF检测方法的功率,该模拟控制了样本量、影响、DIF源和DIF幅度。两种方法都显示出一致的效果,并且在不同的研究条件下都能很好地控制I型误差,这表明已建立的DIF检测方法与MUPP模型一起工作得很好。当DIF源为语句判别时,Lord卡方法优于IPR法;当DIF源为语句阈值时,Lord卡方法优于IPR法。此外,当DIF源为语句位置时,两种方法的性能相似,并且表现出更好的能力,这与先前的研究一致。本文讨论了用MFC检测DIF的研究意义和实际建议,以及局限性。
{"title":"Detecting DIF with the Multi-Unidimensional Pairwise Preference Model: Lord's Chi-square and IPR-NCDIF Methods.","authors":"Lavanya S Kumar, Naidan Tu, Sean Joo, Stephen Stark","doi":"10.1177/01466216251351949","DOIUrl":"10.1177/01466216251351949","url":null,"abstract":"<p><p>Multidimensional forced choice (MFC) measures are gaining prominence in noncognitive assessment. Yet there has been little research on detecting differential item functioning (DIF) with models for forced choice measures. This research extended two well-known DIF detection methods to MFC measures. Specifically, the performance of Lord's chi-square and item parameter replication (IPR) methods with MFC tests based on the Multi-Unidimensional Pairwise Preference (MUPP) model was investigated. The Type I error rate and power of the DIF detection methods were examined in a Monte Carlo simulation that manipulated sample size, impact, DIF source, and DIF magnitude. Both methods showed consistent power and were found to control Type I error well across study conditions, indicating that established approaches to DIF detection work well with the MUPP model. Lord's chi-square outperformed the IPR method when DIF source was statement discrimination while the opposite was true when DIF source was statement threshold. Also, both methods performed similarly and showed better power when DIF source was statement location, in line with previous research. Study implications and practical recommendations for DIF detection with MFC tests, as well as limitations, are discussed.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"01466216251351949"},"PeriodicalIF":1.0,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12213542/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144561576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparing Approaches to Estimating Person Parameters for the MUPP Model. MUPP模型中人参数估计方法的比较。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-01 Epub Date: 2025-01-27 DOI: 10.1177/01466216251316278
David M LaHuis, Caitlin E Blackmore, Gage M Ammons

This study compared maximum a posteriori (MAP), expected a posteriori (EAP), and Markov Chain Monte Carlo (MCMC) approaches to computing person scores from the Multi-Unidimensional Pairwise Preference Model. The MCMC approach used the No-U-Turn sampling (NUTS). Results suggested the EAP with fully crossed quadrature and the NUTS outperformed the others when there were fewer dimensions. In addition, the NUTS produced the most accurate estimates in larger dimension conditions. The number of items per dimension had the largest effect on person parameter recovery.

本研究比较了最大后验(MAP)、期望后验(EAP)和马尔可夫链蒙特卡罗(MCMC)方法在多维成对偏好模型中计算人得分的方法。MCMC方法使用无掉头抽样(NUTS)。结果表明,在低维数情况下,完全交叉正交的EAP和NUTS表现较好。此外,NUTS在较大尺寸条件下产生了最准确的估计。每个维度的项目数对人参数恢复的影响最大。
{"title":"Comparing Approaches to Estimating Person Parameters for the MUPP Model.","authors":"David M LaHuis, Caitlin E Blackmore, Gage M Ammons","doi":"10.1177/01466216251316278","DOIUrl":"10.1177/01466216251316278","url":null,"abstract":"<p><p>This study compared maximum a posteriori (MAP), expected a posteriori (EAP), and Markov Chain Monte Carlo (MCMC) approaches to computing person scores from the Multi-Unidimensional Pairwise Preference Model. The MCMC approach used the No-U-Turn sampling (NUTS). Results suggested the EAP with fully crossed quadrature and the NUTS outperformed the others when there were fewer dimensions. In addition, the NUTS produced the most accurate estimates in larger dimension conditions. The number of items per dimension had the largest effect on person parameter recovery.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"163-176"},"PeriodicalIF":1.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11775930/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143068980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
R Package for Calculating Estimators of the Proportion of Explained Variance and Standardized Regression Coefficients in Multiply Imputed Datasets. R包计算的解释方差和标准化回归系数的比例估计在多重输入数据集。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-01 Epub Date: 2025-01-26 DOI: 10.1177/01466216251316275
Joost R van Ginkel, Julian D Karch
{"title":"R Package for Calculating Estimators of the Proportion of Explained Variance and Standardized Regression Coefficients in Multiply Imputed Datasets.","authors":"Joost R van Ginkel, Julian D Karch","doi":"10.1177/01466216251316275","DOIUrl":"10.1177/01466216251316275","url":null,"abstract":"","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"243-244"},"PeriodicalIF":1.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11770685/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143060765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-Parametric Item Response Theory With O'Sullivan Splines for Item Responses and Response Time. 利用奥沙利文样条对项目响应和响应时间进行半参数项目响应理论研究
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-01 Epub Date: 2025-02-02 DOI: 10.1177/01466216251316277
Chen-Wei Liu

Response time (RT) has been an essential resource for supplementing the estimation accuracy of latent traits and item parameters in educational testing. Most item response theory (IRT) approaches are based on parametric RT models. However, since test takers may alter their behaviors during a test due to motivation or strategy shifts, fatigue, or other causes, parametric IRT models are unlikely to capture such subtle and nonlinear information. In this work, we propose a novel semi-parametric IRT model with O'Sullivan splines to accommodate the flexible mean RT shapes and explore the underlying nonlinear relationships between latent traits and RT. A simulation study was conducted to demonstrate the substantial improvement in parameter estimation achieved by the new model, as well as the detriment of using parametric models in terms of biases and measurement errors. Using this model, a dataset of mathematics test scores and RT from the Programme for International Student Assessment was analyzed to demonstrate the evident nonlinearity and to compare the proposed model with existing models in terms of model fitting. The findings presented in this study indicate the promising nature of the new approach, suggesting its potential as an additional psychometric tool to enhance test reliability and reduce measurement errors.

在教育测试中,反应时间是补充潜在特质和项目参数估计准确性的重要资源。大多数项目反应理论(IRT)方法都是基于参数化反应模型。然而,由于考生在考试过程中可能会由于动机或策略转变、疲劳或其他原因而改变他们的行为,参数IRT模型不太可能捕捉到这种微妙的非线性信息。在这项工作中,我们提出了一种新的半参数IRT模型与O'Sullivan样条,以适应灵活的平均RT形状,并探索潜在性状与RT之间潜在的非线性关系。通过仿真研究,证明了新模型在参数估计方面取得的实质性改进,以及使用参数模型在偏差和测量误差方面的危害。使用该模型,分析了国际学生评估计划的数学考试成绩和RT数据集,以证明明显的非线性,并在模型拟合方面将所提出的模型与现有模型进行比较。这项研究的发现表明了新方法的前景,表明它有可能作为一种额外的心理测量工具来提高测试的可靠性和减少测量误差。
{"title":"Semi-Parametric Item Response Theory With O'Sullivan Splines for Item Responses and Response Time.","authors":"Chen-Wei Liu","doi":"10.1177/01466216251316277","DOIUrl":"10.1177/01466216251316277","url":null,"abstract":"<p><p>Response time (RT) has been an essential resource for supplementing the estimation accuracy of latent traits and item parameters in educational testing. Most item response theory (IRT) approaches are based on parametric RT models. However, since test takers may alter their behaviors during a test due to motivation or strategy shifts, fatigue, or other causes, parametric IRT models are unlikely to capture such subtle and nonlinear information. In this work, we propose a novel semi-parametric IRT model with O'Sullivan splines to accommodate the flexible mean RT shapes and explore the underlying nonlinear relationships between latent traits and RT. A simulation study was conducted to demonstrate the substantial improvement in parameter estimation achieved by the new model, as well as the detriment of using parametric models in terms of biases and measurement errors. Using this model, a dataset of mathematics test scores and RT from the Programme for International Student Assessment was analyzed to demonstrate the evident nonlinearity and to compare the proposed model with existing models in terms of model fitting. The findings presented in this study indicate the promising nature of the new approach, suggesting its potential as an additional psychometric tool to enhance test reliability and reduce measurement errors.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"224-242"},"PeriodicalIF":1.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11789044/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143190883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Compound Optimal Design for Online Item Calibration Under the Two-Parameter Logistic Model. 双参数Logistic模型下在线项目标定的复合优化设计。
IF 1.2 4区 心理学 Q4 PSYCHOLOGY, MATHEMATICAL Pub Date : 2025-07-01 Epub Date: 2025-01-28 DOI: 10.1177/01466216251316276
Lihong Song, Wenyi Wang

Under the theory of sequential design, compound optimal design with two optimality criteria can be used to solve the problem of efficient calibration of item parameters of item response theory model. In order to efficiently calibrate item parameters in computerized testing, a compound optimal design is proposed for the simultaneous estimation of item difficulty and discrimination parameters under the two-parameter logistic model, which adaptively focuses on optimizing the parameter which is difficult to estimate. The compound optimal design using the acceptance probability can provide ability design points to optimize the item difficulty and discrimination parameters, respectively. Simulation and real data analysis studies showed that the compound optimal design outperformed than the D-optimal and random design in terms of the recovery of both discrimination and difficulty parameters.

在序贯设计理论下,采用具有两个最优准则的复合优化设计可解决项目反应理论模型中项目参数的有效标定问题。为了有效地标定计算机测试中的试题参数,提出了一种双参数logistic模型下试题难度和判别参数同时估计的复合优化设计方法,该方法自适应地对难以估计的试题参数进行优化。利用可接受概率的复合优化设计可以分别提供优化项目难度和判别参数的能力设计点。仿真和实际数据分析研究表明,复合优化设计在识别参数和难度参数的恢复方面都优于d -最优设计和随机设计。
{"title":"Compound Optimal Design for Online Item Calibration Under the Two-Parameter Logistic Model.","authors":"Lihong Song, Wenyi Wang","doi":"10.1177/01466216251316276","DOIUrl":"10.1177/01466216251316276","url":null,"abstract":"<p><p>Under the theory of sequential design, compound optimal design with two optimality criteria can be used to solve the problem of efficient calibration of item parameters of item response theory model. In order to efficiently calibrate item parameters in computerized testing, a compound optimal design is proposed for the simultaneous estimation of item difficulty and discrimination parameters under the two-parameter logistic model, which adaptively focuses on optimizing the parameter which is difficult to estimate. The compound optimal design using the acceptance probability can provide ability design points to optimize the item difficulty and discrimination parameters, respectively. Simulation and real data analysis studies showed that the compound optimal design outperformed than the D-optimal and random design in terms of the recovery of both discrimination and difficulty parameters.</p>","PeriodicalId":48300,"journal":{"name":"Applied Psychological Measurement","volume":" ","pages":"177-198"},"PeriodicalIF":1.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11775943/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143068983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied Psychological Measurement
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1