首页 > 最新文献

Educational and Psychological Measurement最新文献

英文 中文
Faking in High-Stakes Personality Assessments: A Response-Time-Based Latent Response Mixture Modeling Approach. 高风险人格评估中的作假:一种基于反应时间的潜在反应混合建模方法。
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-03-18 DOI: 10.1177/00131644261422169
Timo Seitz, Esther Ulitzsch

When personality assessments are employed in high-stakes contexts, there is the risk that test-takers provide overly positive descriptions of themselves. This response bias is known as faking and has often been addressed in latent variable models through an additional dimension capturing each test-taker's faking degree. Such models typically assume a homogeneous response strategy for all test-takers, with substantive traits and faking jointly influencing responses to all items. In this article, we present a latent response mixture item response theory (IRT) model of faking that accounts for changes in test-takers' response strategies over the course of the assessment. The model translates theoretical considerations about test-taker behavior into different model components for item responses and corresponding item-level response times (RT), thereby allowing to account for, identify, and investigate different faking-related response strategies on the person-by-item level. In a parameter recovery study, we found that the model parameters can be estimated well under realistic conditions. Also, we applied the model to an empirical dataset (N = 1,824) from a job application context, showcasing its utility in real high-stakes assessment data. We conclude the article by discussing the role of the model for psychological measurement as well as substantive research.

当在高风险环境中进行性格评估时,有可能出现考生对自己的描述过于积极的情况。这种反应偏差被称为伪造,通常在潜在变量模型中通过捕获每个考生伪造程度的额外维度来解决。这些模型通常假设所有考生的反应策略都是同质的,实质性特征和虚假特征共同影响对所有项目的反应。在本文中,我们提出了一个潜在反应混合项目反应理论(IRT)模型,该模型解释了考生在评估过程中反应策略的变化。该模型将有关考生行为的理论考虑转化为项目反应和相应的项目级反应时间(RT)的不同模型组件,从而允许在个人-项目水平上解释、识别和研究不同的假相关反应策略。在参数恢复研究中,我们发现模型参数在实际条件下可以很好地估计。此外,我们将该模型应用于工作申请背景下的经验数据集(N = 1,824),以展示其在真实高风险评估数据中的实用性。最后,我们讨论了该模型在心理测量中的作用以及实质性的研究。
{"title":"Faking in High-Stakes Personality Assessments: A Response-Time-Based Latent Response Mixture Modeling Approach.","authors":"Timo Seitz, Esther Ulitzsch","doi":"10.1177/00131644261422169","DOIUrl":"10.1177/00131644261422169","url":null,"abstract":"<p><p>When personality assessments are employed in high-stakes contexts, there is the risk that test-takers provide overly positive descriptions of themselves. This response bias is known as faking and has often been addressed in latent variable models through an additional dimension capturing each test-taker's faking degree. Such models typically assume a homogeneous response strategy for all test-takers, with substantive traits and faking jointly influencing responses to all items. In this article, we present a latent response mixture item response theory (IRT) model of faking that accounts for changes in test-takers' response strategies over the course of the assessment. The model translates theoretical considerations about test-taker behavior into different model components for item responses and corresponding item-level response times (RT), thereby allowing to account for, identify, and investigate different faking-related response strategies on the person-by-item level. In a parameter recovery study, we found that the model parameters can be estimated well under realistic conditions. Also, we applied the model to an empirical dataset (<i>N</i> = 1,824) from a job application context, showcasing its utility in real high-stakes assessment data. We conclude the article by discussing the role of the model for psychological measurement as well as substantive research.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644261422169"},"PeriodicalIF":2.3,"publicationDate":"2026-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12999537/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147497889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Conditional Dependencies Between Response Time and Item Discrimination: An Item-Level Meta-Analysis. 反应时间与项目辨别力的条件依赖关系:一个项目水平的元分析。
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-03-17 DOI: 10.1177/00131644261426972
Joshua B Gilbert, William S Young, Zachary Himmelsbach, Esther Ulitzsch, Benjamin W Domingue

The use of process data, such as response time (RT) in psychometrics, has generally focused on the relationship between speed and accuracy. The potential relationships between RT and item discrimination remain less explored. In this study, we propose a model for simultaneously estimating the relationships between RT and item discrimination at the person, item, and person-by-item (residual) levels and illustrate our approach through an item-level meta-analysis of 40 empirical data sets comprising 1.84 million item responses. We find no evidence of average differences in item discrimination between items of different time intensity or persons of different average RT, while residual RT strongly and negatively predicts item discrimination (pooled coef. = -.27% per 1% difference in RT, SE = .04, τ = .17). While heterogeneity is high, we find little evidence of moderation by overall data set characteristics. Flexible generalized additive models show that the relationship between residual RT and item discrimination is generally curvilinear, with discrimination maximized just below average RT and minimized at the extremes. Our results suggest that RT data can provide insights into the measurement properties of educational and psychological assessments, but that the relationships between RT and item discrimination are highly variable.

过程数据的使用,如心理测量学中的反应时间(RT),通常集中在速度和准确性之间的关系上。RT和项目歧视之间的潜在关系仍然很少被探索。在这项研究中,我们提出了一个模型来同时估计RT和项目歧视在人、项目和人对项目(残差)水平之间的关系,并通过包含184万个项目回答的40个经验数据集的项目水平荟萃分析来说明我们的方法。我们发现不同时间强度的项目或不同平均RT的人在项目歧视上没有平均差异的证据,而残差RT对项目歧视有显著的负向预测(汇总系数)。= - 0.27% / 1%的RT差异,SE = 0.04, τ = 0.17)。虽然异质性很高,但我们发现总体数据集特征的适度证据很少。灵活的广义加性模型表明,残差RT与项目识别率之间的关系一般呈曲线关系,识别率在低于平均RT时最大,在极端情况下最小。我们的研究结果表明,RT数据可以深入了解教育和心理评估的测量特性,但RT与项目歧视之间的关系是高度可变的。
{"title":"Conditional Dependencies Between Response Time and Item Discrimination: An Item-Level Meta-Analysis.","authors":"Joshua B Gilbert, William S Young, Zachary Himmelsbach, Esther Ulitzsch, Benjamin W Domingue","doi":"10.1177/00131644261426972","DOIUrl":"https://doi.org/10.1177/00131644261426972","url":null,"abstract":"<p><p>The use of process data, such as response time (RT) in psychometrics, has generally focused on the relationship between speed and accuracy. The potential relationships between RT and item discrimination remain less explored. In this study, we propose a model for simultaneously estimating the relationships between RT and item discrimination at the person, item, and person-by-item (residual) levels and illustrate our approach through an item-level meta-analysis of 40 empirical data sets comprising 1.84 million item responses. We find no evidence of average differences in item discrimination between items of different time intensity or persons of different average RT, while residual RT strongly and negatively predicts item discrimination (pooled coef. = -.27% per 1% difference in RT, <i>SE</i> = .04, <math><mrow><mi>τ</mi></mrow> </math> = .17). While heterogeneity is high, we find little evidence of moderation by overall data set characteristics. Flexible generalized additive models show that the relationship between residual RT and item discrimination is generally curvilinear, with discrimination maximized just below average RT and minimized at the extremes. Our results suggest that RT data can provide insights into the measurement properties of educational and psychological assessments, but that the relationships between RT and item discrimination are highly variable.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644261426972"},"PeriodicalIF":2.3,"publicationDate":"2026-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12995739/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147485016","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimating Trends With Differential Item Functioning: A Comparison of Five IRT-Based Approaches. 用差异项目功能估计趋势:五种基于红外光谱的方法的比较。
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-03-13 DOI: 10.1177/00131644251408818
Oskar Engels, Oliver Lüdtke, Alexander Robitzsch

In longitudinal assessments, tests are frequently used to estimate trends over time. However, when item parameters lack invariance, time-point comparisons can be distorted, necessitating appropriate statistical methods to achieve accurate estimation. This study compares trend estimates using the two-parameter logistic (2PL) model under item parameter drift (IPD) across five trend-estimation approaches for two time points: First, concurrent calibration, which jointly estimates item parameters across multiple time points. Second, fixed calibration, which estimates item parameters at a single time point and fixes them at the other time point. Third, robust linking with Haberman and Haebara as linking methods with L p or L 0 losses. Fourth, non-invariant items are detected using likelihood-ratio tests or the root mean square deviation statistic with fixed or data-driven cutoffs, and trend estimates are then recomputed using only the detected invariant items under partial invariance. Fifth, regularized estimation under a smooth Bayesian information criterion (SBIC) is applied, shrinking small or null IPD effects toward zero while estimating all others as nonzero. Bias and relative root mean square error (RMSE) were evaluated for the mean and SD at T2. An empirical example using synthetic longitudinal reading data, applying the trend-estimation approaches, is provided. The results indicate that the regularized estimation with SBIC performed best across conditions, maintaining low bias and RMSE, followed by robust linking methods. Specifically, Haberman linking with the L 0 loss function showed superior performance under unbalanced IPD, outperforming the partial invariance approaches. Concurrent and fixed calibration showed the poorest trend recovery under unbalanced IPD conditions.

在纵向评估中,测试经常用于估计一段时间内的趋势。然而,当项目参数缺乏不变性时,时间点比较可能会失真,需要适当的统计方法来实现准确的估计。本研究比较了项目参数漂移(IPD)下的两参数logistic (2PL)模型在五个时间点趋势估计方法下的趋势估计:第一,同步校准,联合估计多个时间点的项目参数。第二种是固定校准,即在一个时间点估计项目参数,并在另一个时间点固定它们。第三,利用Haberman和Haebara作为lp或l0损失的连接方法进行鲁棒连接。第四,使用似然比检验或具有固定或数据驱动截止点的均方根偏差统计来检测非不变项,然后仅使用在部分不变性下检测到的不变项重新计算趋势估计。第五,应用平滑贝叶斯信息准则(SBIC)下的正则化估计,将小的或零的IPD效应向零缩小,同时将所有其他的IPD效应估计为非零。对T2时的平均值和SD进行偏倚和相对均方根误差(RMSE)评估。最后给出了一个应用趋势估计方法的综合纵向读数数据的实例。结果表明,基于SBIC的正则化估计在不同条件下表现最佳,保持低偏差和RMSE,其次是鲁棒连接方法。具体而言,Haberman与l0损失函数的连接在不平衡IPD下表现出优异的性能,优于部分不变性方法。在不平衡IPD条件下,同步校准和固定校准的趋势恢复最差。
{"title":"Estimating Trends With Differential Item Functioning: A Comparison of Five IRT-Based Approaches.","authors":"Oskar Engels, Oliver Lüdtke, Alexander Robitzsch","doi":"10.1177/00131644251408818","DOIUrl":"https://doi.org/10.1177/00131644251408818","url":null,"abstract":"<p><p>In longitudinal assessments, tests are frequently used to estimate trends over time. However, when item parameters lack invariance, time-point comparisons can be distorted, necessitating appropriate statistical methods to achieve accurate estimation. This study compares trend estimates using the two-parameter logistic (2PL) model under item parameter drift (IPD) across five trend-estimation approaches for two time points: First, concurrent calibration, which jointly estimates item parameters across multiple time points. Second, fixed calibration, which estimates item parameters at a single time point and fixes them at the other time point. Third, robust linking with Haberman and Haebara as linking methods with <math> <mrow> <msub><mrow><mi>L</mi></mrow> <mrow><mi>p</mi></mrow> </msub> </mrow> </math> or <math> <mrow> <msub><mrow><mi>L</mi></mrow> <mrow><mn>0</mn></mrow> </msub> </mrow> </math> losses. Fourth, non-invariant items are detected using likelihood-ratio tests or the root mean square deviation statistic with fixed or data-driven cutoffs, and trend estimates are then recomputed using only the detected invariant items under partial invariance. Fifth, regularized estimation under a smooth Bayesian information criterion (SBIC) is applied, shrinking small or null IPD effects toward zero while estimating all others as nonzero. Bias and relative root mean square error (RMSE) were evaluated for the mean and <i>SD</i> at T2. An empirical example using synthetic longitudinal reading data, applying the trend-estimation approaches, is provided. The results indicate that the regularized estimation with SBIC performed best across conditions, maintaining low bias and RMSE, followed by robust linking methods. Specifically, Haberman linking with the <math> <mrow> <msub><mrow><mi>L</mi></mrow> <mrow><mn>0</mn></mrow> </msub> </mrow> </math> loss function showed superior performance under unbalanced IPD, outperforming the partial invariance approaches. Concurrent and fixed calibration showed the poorest trend recovery under unbalanced IPD conditions.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644251408818"},"PeriodicalIF":2.3,"publicationDate":"2026-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12987755/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147462744","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Discriminating Between Attribute, Item-Position, and Wording Effects by the Congeneric and Tau-Equivalent Confirmatory Factor Analysis Models. 用同类和等效验证性因子分析模型区分属性、项目位置和措辞效应。
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-03-11 DOI: 10.1177/00131644261419028
Karl Schweizer, Xuezhu Ren, Tengfei Wang

The capability of confirmatory factor analysis to discriminate common systematic variation of attribute, item-position, and wording effects was investigated using the congeneric and tau-equivalent models. The simulated data generated according to four approaches included gradually increased amounts of item-position or wording effect variation while the amount of attribute variation was kept constant. The congeneric model always signified good model fit independently of the type and amount of additional common systematic variation, that is, there was no discrimination. In applications of the tau-equivalent model, the increase of the item-position or wording effect variation led to the change from indicating good fit to bad model fit, that is, there was negative discrimination. In contrast, the additionally considered two-factor tau model discriminated positively. As a consequence of these results, we recommend the pre-screening of data for method effects.

验证性因子分析的能力,以区分常见的系统变化的属性,项目位置和措辞的影响,采用同属和对等模型。四种方法生成的模拟数据包括:物品位置或措辞效果变化量逐渐增加,而属性变化量保持不变。无论附加的共同系统变异的类型和数量如何,相同的模型总是具有良好的模型拟合,即不存在判别。在tau等效模型的应用中,项目位置或措辞效应变异的增加导致模型拟合从良好变为不良,即存在负判别。相比之下,另外考虑的双因素tau模型是积极的。由于这些结果,我们建议预先筛选数据的方法效果。
{"title":"Discriminating Between Attribute, Item-Position, and Wording Effects by the Congeneric and Tau-Equivalent Confirmatory Factor Analysis Models.","authors":"Karl Schweizer, Xuezhu Ren, Tengfei Wang","doi":"10.1177/00131644261419028","DOIUrl":"https://doi.org/10.1177/00131644261419028","url":null,"abstract":"<p><p>The capability of confirmatory factor analysis to discriminate common systematic variation of attribute, item-position, and wording effects was investigated using the congeneric and tau-equivalent models. The simulated data generated according to four approaches included gradually increased amounts of item-position or wording effect variation while the amount of attribute variation was kept constant. The congeneric model always signified good model fit independently of the type and amount of additional common systematic variation, that is, there was no discrimination. In applications of the tau-equivalent model, the increase of the item-position or wording effect variation led to the change from indicating good fit to bad model fit, that is, there was negative discrimination. In contrast, the additionally considered two-factor tau model discriminated positively. As a consequence of these results, we recommend the pre-screening of data for method effects.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644261419028"},"PeriodicalIF":2.3,"publicationDate":"2026-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12979218/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147462748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimation of Conditional Standard Errors of Measurement for MLE Scores in MST. MST中MLE分数的条件标准误差估计。
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-02-25 DOI: 10.1177/00131644261420391
Yuanyuan J Stirn, Won-Chan Lee

This paper proposes an information-based analytic method for calculating the conditional standard error of measurement (CSEM) in multistage testing (MST) using maximum likelihood estimation. The accuracy of the proposed method was evaluated by comparing CSEMs computed using the analytic method with those obtained from simulation across the same four MST designs. The results show that analytic and simulation-based CSEMs converge as test length increases, indicating that the proposed method provides a reliable approximation for longer tests. However, shorter tests and more complex MST designs require additional items to achieve comparable accuracy. The study also compared the proposed method with Park et al.'s analytic approach. Practical implications of the proposed method are discussed.

提出了一种利用极大似然估计计算多阶段测试中条件测量标准误差的信息分析方法。通过比较分析方法计算的csem与模拟得到的相同四种MST设计的csem,评估了所提方法的准确性。结果表明,随着试验长度的增加,基于解析和仿真的cems收敛,表明该方法为较长的试验提供了可靠的近似方法。然而,更短的测试和更复杂的MST设计需要额外的项目来达到相当的准确性。该研究还将提出的方法与Park等人的分析方法进行了比较。讨论了该方法的实际意义。
{"title":"Estimation of Conditional Standard Errors of Measurement for MLE Scores in MST.","authors":"Yuanyuan J Stirn, Won-Chan Lee","doi":"10.1177/00131644261420391","DOIUrl":"https://doi.org/10.1177/00131644261420391","url":null,"abstract":"<p><p>This paper proposes an information-based analytic method for calculating the conditional standard error of measurement (CSEM) in multistage testing (MST) using maximum likelihood estimation. The accuracy of the proposed method was evaluated by comparing CSEMs computed using the analytic method with those obtained from simulation across the same four MST designs. The results show that analytic and simulation-based CSEMs converge as test length increases, indicating that the proposed method provides a reliable approximation for longer tests. However, shorter tests and more complex MST designs require additional items to achieve comparable accuracy. The study also compared the proposed method with Park et al.'s analytic approach. Practical implications of the proposed method are discussed.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644261420391"},"PeriodicalIF":2.3,"publicationDate":"2026-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12945742/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147324610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Misclassification Produced by Rapid-Guessing Identification Methods and Their Suitability Under Various Conditions. 快速猜测识别方法产生的误分类及其在不同条件下的适用性。
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-02-23 DOI: 10.1177/00131644261419426
Santeri Holopainen, Jari Metsämuuronen, Mikko-Jussi Laakso, Janne Kujala

Response Time Threshold Methods (RTTMs) are widely used to identify rapid-guessing behavior (RG) in low-stakes assessments, yet face two key challenges: (a) inevitable misclassifications due to overlapping response time distributions of engaged and disengaged responses, and (b) lack of agreement on which method to use under varying conditions. This simulation study evaluated five RTTMs. Item responses and response times were generated from either a one-component model without RG or a two-component mixture model with RG in the population. Distribution, item, and person parameters were varied. Results showed that when the population contained RG, the mixture lognormal distribution-based method (MLN) was the most robust approach and estimated precise thresholds closest to the time points at which the misclassification rates were minimized, even when bimodality was more difficult to detect. The cumulative proportion method (CUMP) was less robust but also accurate when successful, though less precise. In addition, when the population did not include RG, CUMP was the only method to set thresholds for a notable proportion of cases. The methods were generally more conservative than liberal, though the mixture response time quantile method (MRTQ) was neither. The results are discussed in the light of prior RG research and the methods' characteristics, and future directions are suggested. Ultimately, for practical settings, we recommend a six-step process for RG identification that utilizes both a mixture modeling approach (MLN or MRTQ) and the CUMP method.

反应时间阈值方法(RTTMs)被广泛用于识别低风险评估中的快速猜测行为(RG),但面临两个关键挑战:(a)由于参与和不参与反应的反应时间分布重叠而不可避免的错误分类;(b)在不同条件下使用哪种方法缺乏共识。该模拟研究评估了5种rttm。项目反应和反应时间由不含RG的单组分模型或含RG的双组分混合模型生成。分布、项目和人员参数各不相同。结果表明,当总体包含RG时,基于混合对数正态分布的方法(MLN)是最稳健的方法,即使在双峰更难检测的情况下,它也能准确估计出最接近误分类率最小的时间点的阈值。累积比例法(CUMP)的稳健性较差,但在成功时也很准确,尽管精度较低。此外,当人群不包括RG时,CUMP是为显著比例的病例设置阈值的唯一方法。虽然混合反应时间分位数法(MRTQ)两者均不保守,但该方法总体上偏保守。结合前人的研究成果和方法特点,对研究结果进行了讨论,并提出了今后的研究方向。最后,对于实际设置,我们推荐使用混合建模方法(MLN或MRTQ)和CUMP方法进行RG识别的六步过程。
{"title":"Misclassification Produced by Rapid-Guessing Identification Methods and Their Suitability Under Various Conditions.","authors":"Santeri Holopainen, Jari Metsämuuronen, Mikko-Jussi Laakso, Janne Kujala","doi":"10.1177/00131644261419426","DOIUrl":"https://doi.org/10.1177/00131644261419426","url":null,"abstract":"<p><p>Response Time Threshold Methods (RTTMs) are widely used to identify rapid-guessing behavior (RG) in low-stakes assessments, yet face two key challenges: (a) inevitable misclassifications due to overlapping response time distributions of engaged and disengaged responses, and (b) lack of agreement on which method to use under varying conditions. This simulation study evaluated five RTTMs. Item responses and response times were generated from either a one-component model without RG or a two-component mixture model with RG in the population. Distribution, item, and person parameters were varied. Results showed that when the population contained RG, the mixture lognormal distribution-based method (MLN) was the most robust approach and estimated precise thresholds closest to the time points at which the misclassification rates were minimized, even when bimodality was more difficult to detect. The cumulative proportion method (CUMP) was less robust but also accurate when successful, though less precise. In addition, when the population did not include RG, CUMP was the only method to set thresholds for a notable proportion of cases. The methods were generally more conservative than liberal, though the mixture response time quantile method (MRTQ) was neither. The results are discussed in the light of prior RG research and the methods' characteristics, and future directions are suggested. Ultimately, for practical settings, we recommend a six-step process for RG identification that utilizes both a mixture modeling approach (MLN or MRTQ) and the CUMP method.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644261419426"},"PeriodicalIF":2.3,"publicationDate":"2026-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12929091/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147303530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From Agreement to Epistemic Alignment: A Signal Detection-Theoretic Model of Inter-Rater Reliability. 从一致性到认知一致性:一个信号检测理论模型。
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-02-16 DOI: 10.1177/00131644261417643
Irene Gianeselli

Inter-rater reliability is commonly assessed using chance-corrected agreement coefficients such as Cohen's κ, which summarize concordance among categorical judgments without modeling the inferential processes that generate them. As a result, κ is sensitive to prevalence imbalance, task difficulty, and heterogeneity in decision criteria and is often misinterpreted as a proxy for diagnostic accuracy or rater competence. This paper reframes inter-rater reliability within a signal detection-theoretic (SDT) framework in which categorical judgments arise from comparisons between latent continuous evidence and rater-specific decision thresholds. Within this generative model, κ can be interpreted as a bounded transformation of discrete strategic variance (i.e., the observable consequence of dispersion in latent decision criteria) rather than as a direct measure of epistemic alignment. To make this structure explicit, we introduce the Strategic Convergence Index (SCI), a normalized functional summarizing convergence in rater decision thresholds under an SDT generative process. SCI is not proposed as a standalone agreement coefficient but as a model-implied quantity whose interpretation depends on explicit assumptions about evidence distributions and decision rules. Monte Carlo simulations show that κ varies systematically with prevalence and perceptual discriminability even when decision-policy alignment is held constant, whereas SCI selectively tracks epistemic alignment and remains invariant to these factors. Supplementary model-based analyses further illustrate that SCI can be recovered as a stable system-level property even under latent-truth uncertainty, whereas individual thresholds may be weakly identified. Together, these results clarify the epistemic meaning of κ and motivate a decomposition of inter-rater reliability into outcome-level agreement and process-level alignment. By linking classical agreement statistics to an explicit generative model of judgment, the Strategic Convergence framework advances reliability assessment from description toward explanation.

评级者之间的可靠性通常使用机会校正的一致系数(如Cohen’s κ)来评估,该系数总结了分类判断之间的一致性,而无需对产生这些判断的推理过程进行建模。因此,κ对患病率不平衡、任务难度和决策标准的异质性很敏感,经常被误解为诊断准确性或更高能力的代表。本文在信号检测理论(SDT)框架内重新构建了评级间的可靠性,其中类别判断来自潜在连续证据和评级特定决策阈值之间的比较。在这个生成模型中,κ可以被解释为离散策略方差的有界变换(即,潜在决策标准中分散的可观察结果),而不是作为认知一致性的直接度量。为了明确这一结构,我们引入了策略收敛指数(SCI),这是一种标准化的函数,总结了SDT生成过程中不同决策阈值的收敛性。SCI不是作为一个独立的一致系数,而是作为一个模型隐含量,其解释依赖于对证据分布和决策规则的明确假设。蒙特卡罗模拟表明,即使在决策-政策一致性保持不变的情况下,κ也会随着普遍性和感知可辨性而系统地变化,而SCI则选择性地跟踪认知一致性,并对这些因素保持不变。补充的基于模型的分析进一步表明,即使在潜在真值不确定性下,SCI也可以作为稳定的系统级属性恢复,而个体阈值可能被弱识别。总之,这些结果阐明了κ的认知意义,并激发了将评分者之间的可靠性分解为结果级一致性和过程级一致性。通过将经典协议统计与明确的判断生成模型联系起来,战略收敛框架将可靠性评估从描述推进到解释。
{"title":"From Agreement to Epistemic Alignment: A Signal Detection-Theoretic Model of Inter-Rater Reliability.","authors":"Irene Gianeselli","doi":"10.1177/00131644261417643","DOIUrl":"https://doi.org/10.1177/00131644261417643","url":null,"abstract":"<p><p>Inter-rater reliability is commonly assessed using chance-corrected agreement coefficients such as Cohen's κ, which summarize concordance among categorical judgments without modeling the inferential processes that generate them. As a result, κ is sensitive to prevalence imbalance, task difficulty, and heterogeneity in decision criteria and is often misinterpreted as a proxy for diagnostic accuracy or rater competence. This paper reframes inter-rater reliability within a signal detection-theoretic (SDT) framework in which categorical judgments arise from comparisons between latent continuous evidence and rater-specific decision thresholds. Within this generative model, κ can be interpreted as a bounded transformation of discrete strategic variance (i.e., the observable consequence of dispersion in latent decision criteria) rather than as a direct measure of epistemic alignment. To make this structure explicit, we introduce the Strategic Convergence Index (SCI), a normalized functional summarizing convergence in rater decision thresholds under an SDT generative process. SCI is not proposed as a standalone agreement coefficient but as a model-implied quantity whose interpretation depends on explicit assumptions about evidence distributions and decision rules. Monte Carlo simulations show that κ varies systematically with prevalence and perceptual discriminability even when decision-policy alignment is held constant, whereas SCI selectively tracks epistemic alignment and remains invariant to these factors. Supplementary model-based analyses further illustrate that SCI can be recovered as a stable system-level property even under latent-truth uncertainty, whereas individual thresholds may be weakly identified. Together, these results clarify the epistemic meaning of κ and motivate a decomposition of inter-rater reliability into outcome-level agreement and process-level alignment. By linking classical agreement statistics to an explicit generative model of judgment, the Strategic Convergence framework advances reliability assessment from description toward explanation.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644261417643"},"PeriodicalIF":2.3,"publicationDate":"2026-02-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12909152/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146219104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Consistency of Automatic Scoring with Large Language Models. 基于大语言模型的自动评分一致性研究。
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-02-16 DOI: 10.1177/00131644261418138
Mingfeng Xue, Xingyao Xiao, Yunting Liu, Mark Wilson

Large language models (LLMs) have shown great potential in automatic scoring. However, due to model characteristics and variation in training materials and pipelines, scoring inconsistency can exist within an LLM and across LLMs when rating the same response multiple times. This study investigates the intra-LLM and inter-LLM consistency in scoring with five LLMs (i.e., Claude, DeepSeek, Gemini, GPT, and Qwen), variability under different temperatures, and their relationship with scoring accuracy. Moreover, a voting strategy that assembles information from different LLMs was proposed to address inconsistent scoring. Using constructed-response items from a science education assessment and open-source data from the Automated Student Assessment Prize (ASAP), we find that: (a) LLMs generally exhibited almost perfect intra-LLM consistency regardless of temperature; (b) inter-LLM consistency was moderate, with higher agreement observed for items that were easier to score; (c) intra-LLM consistency consistently exceeded inter-LLM consistency, supporting the expectation that within-model consistency represents an upper bound for cross-model agreement; (d) intra-LLM consistency was not associated with scoring accuracy, whereas inter-LLM consistency showed a strong positive relationship with accuracy; and (e) majority voting across LLMs improved scoring accuracy by leveraging complementary strengths of different models.

大型语言模型(llm)在自动评分方面显示出巨大的潜力。然而,由于模型特征和培训材料和管道的变化,在多次对相同的响应进行评分时,评分不一致可能存在于LLM内部和LLM之间。本研究探讨了五个llm (Claude, DeepSeek, Gemini, GPT, Qwen)在llm内部和llm之间评分的一致性,不同温度下的变异性及其与评分准确性的关系。此外,提出了一种整合不同llm信息的投票策略,以解决评分不一致的问题。使用科学教育评估中的构建反应项目和自动化学生评估奖(ASAP)的开源数据,我们发现:(a)无论温度如何,llm通常表现出几乎完美的llm内部一致性;(b) llm之间的一致性是中等的,在更容易评分的项目上观察到更高的一致性;(c) llm内一致性始终超过llm间一致性,支持模型内一致性代表跨模型一致性上界的预期;(d) llm内部一致性与评分准确性不相关,而llm之间的一致性与准确性呈强正相关;(e)通过利用不同模型的互补优势,跨llm的多数投票提高了评分准确性。
{"title":"On the Consistency of Automatic Scoring with Large Language Models.","authors":"Mingfeng Xue, Xingyao Xiao, Yunting Liu, Mark Wilson","doi":"10.1177/00131644261418138","DOIUrl":"https://doi.org/10.1177/00131644261418138","url":null,"abstract":"<p><p>Large language models (LLMs) have shown great potential in automatic scoring. However, due to model characteristics and variation in training materials and pipelines, scoring inconsistency can exist within an LLM and across LLMs when rating the same response multiple times. This study investigates the intra-LLM and inter-LLM consistency in scoring with five LLMs (i.e., Claude, DeepSeek, Gemini, GPT, and Qwen), variability under different temperatures, and their relationship with scoring accuracy. Moreover, a voting strategy that assembles information from different LLMs was proposed to address inconsistent scoring. Using constructed-response items from a science education assessment and open-source data from the Automated Student Assessment Prize (ASAP), we find that: (a) LLMs generally exhibited almost perfect intra-LLM consistency regardless of temperature; (b) inter-LLM consistency was moderate, with higher agreement observed for items that were easier to score; (c) intra-LLM consistency consistently exceeded inter-LLM consistency, supporting the expectation that within-model consistency represents an upper bound for cross-model agreement; (d) intra-LLM consistency was not associated with scoring accuracy, whereas inter-LLM consistency showed a strong positive relationship with accuracy; and (e) majority voting across LLMs improved scoring accuracy by leveraging complementary strengths of different models.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644261418138"},"PeriodicalIF":2.3,"publicationDate":"2026-02-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12909151/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146219113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparing Different Approaches of (Not) Accounting for Rapid Guessing in Plausible Values Estimation. 似是值估计中快速猜测的不同(非)会计方法比较。
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-13 DOI: 10.1177/00131644251395590
Jana Welling, Eva Zink, Timo Gnambs

Educational large-scale assessments provide information on ability differences between groups, informing policies and shaping educational decisions. However, some of these differences might partly reflect variations in test-taking motivation rather than in actual abilities. Existing approaches for mitigating the distorting effects of rapid guessing focus mainly on point estimates of abilities, although research questions often refer to latent variables. The present study seeks to (a) determine the bias introduced by rapid guessing in group comparisons based on plausible value estimates and (b) introduce and evaluate different approaches of handling rapid guessing in the estimation of plausible values. In a simulation study, four models were compared: (1) a baseline model did not account for rapid guessing, (2) a person-level model incorporated rapid guessing as a respondent characteristic in the background model, (3) a response-level model filtered responses with item response times lower than a predetermined threshold, and (4) a combined model merged the person- and response-level approaches. Results show that the response-level and combined model performed best while accounting for rapid guessing on the person level did not suffice. An empirical example using data from a German large-scale assessment (N = 478) demonstrates the applicability of all approaches in practice. Recommendations for future research are given to improve ability estimation.

教育大规模评估提供了群体之间能力差异的信息,为政策提供信息并形成教育决策。然而,其中一些差异可能部分反映了考试动机的差异,而不是实际能力的差异。现有的减轻快速猜测的扭曲效应的方法主要集中在能力的点估计上,尽管研究问题经常涉及潜在变量。本研究旨在(a)确定基于可信值估计的快速猜测在群体比较中引入的偏差,(b)介绍和评估在估计可信值时处理快速猜测的不同方法。在模拟研究中,对四种模型进行了比较:(1)基线模型不考虑快速猜测;(2)个人水平模型将快速猜测作为被调查者的特征纳入背景模型;(3)反应水平模型过滤了项目反应时间低于预定阈值的反应;(4)综合模型将个人和反应水平方法合并。结果表明,反应水平和组合模型表现最好,而考虑到个人水平的快速猜测是不够的。一个使用德国大规模评估(N = 478)数据的实证例子证明了所有方法在实践中的适用性。对今后的研究提出了改进能力估计的建议。
{"title":"Comparing Different Approaches of (Not) Accounting for Rapid Guessing in Plausible Values Estimation.","authors":"Jana Welling, Eva Zink, Timo Gnambs","doi":"10.1177/00131644251395590","DOIUrl":"10.1177/00131644251395590","url":null,"abstract":"<p><p>Educational large-scale assessments provide information on ability differences between groups, informing policies and shaping educational decisions. However, some of these differences might partly reflect variations in test-taking motivation rather than in actual abilities. Existing approaches for mitigating the distorting effects of rapid guessing focus mainly on point estimates of abilities, although research questions often refer to latent variables. The present study seeks to (a) determine the bias introduced by rapid guessing in group comparisons based on plausible value estimates and (b) introduce and evaluate different approaches of handling rapid guessing in the estimation of plausible values. In a simulation study, four models were compared: (1) a baseline model did not account for rapid guessing, (2) a person-level model incorporated rapid guessing as a respondent characteristic in the background model, (3) a response-level model filtered responses with item response times lower than a predetermined threshold, and (4) a combined model merged the person- and response-level approaches. Results show that the response-level and combined model performed best while accounting for rapid guessing on the person level did not suffice. An empirical example using data from a German large-scale assessment (<i>N</i> = 478) demonstrates the applicability of all approaches in practice. Recommendations for future research are given to improve ability estimation.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644251395590"},"PeriodicalIF":2.3,"publicationDate":"2026-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12804065/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145997547","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Consistent Factor Score Regression: A Better Alternative for Uncorrected Factor Score Regression? 一致性因子得分回归:未校正因子得分回归的更好选择?
IF 2.3 3区 心理学 Q2 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-04 DOI: 10.1177/00131644251399588
Jasper Bogaert, Wen Wei Loh, Yves Rosseel

Researchers in the behavioral, educational, and social sciences often aim to analyze relationships among latent variables. Structural equation modeling (SEM) is widely regarded as the gold standard for this purpose. A straightforward alternative for estimating the structural model parameters is uncorrected factor score regression (UFSR), where factor scores are first computed and then employed in regression or path analysis. Unfortunately, the most commonly used factor scores (i.e., Regression and Bartlett factor scores) may yield biased estimates and invalid inferences when using this approach. In recent years, factor score regression (FSR) has enjoyed several methodological advancements to address this inconsistency. Despite these advancements, the use of FSR with correlation-preserving factor scores, here termed consistent factor score regression (cFSR), has received limited attention. In this paper, we revisit cFSR and compare its advantages and disadvantages relative to other recent FSR and SEM methods. We conducted an extensive simulation study comparing cFSR with other estimation approaches, assessing their performance in terms of convergence rate, bias, efficiency, and type I error rate. The findings indicate that cFSR outperforms UFSR while maintaining the conceptual simplicity of UFSR. We encourage behavioral, educational, and social science researchers to avoid UFSR and adopt cFSR as an alternative to SEM.

行为科学、教育科学和社会科学的研究人员经常致力于分析潜在变量之间的关系。结构方程模型(SEM)被广泛认为是这方面的金标准。估计结构模型参数的直接替代方法是未校正因子得分回归(UFSR),其中首先计算因子得分,然后将其用于回归或路径分析。不幸的是,当使用这种方法时,最常用的因子得分(即回归和巴特利特因子得分)可能会产生有偏差的估计和无效的推断。近年来,因子得分回归(FSR)在解决这种不一致性方面取得了一些方法上的进步。尽管取得了这些进步,但将FSR与保持相关性的因子得分(这里称为一致因子得分回归(cFSR))结合使用,受到的关注有限。在本文中,我们回顾了cFSR,并比较了它相对于其他最近的FSR和SEM方法的优点和缺点。我们进行了广泛的模拟研究,比较了cFSR和其他估计方法,评估了它们在收敛速度、偏差、效率和I型错误率方面的性能。研究结果表明,cFSR优于UFSR,同时保持了UFSR概念的简单性。我们鼓励行为科学、教育科学和社会科学研究人员避免UFSR,而采用cFSR作为SEM的替代品。
{"title":"Consistent Factor Score Regression: A Better Alternative for Uncorrected Factor Score Regression?","authors":"Jasper Bogaert, Wen Wei Loh, Yves Rosseel","doi":"10.1177/00131644251399588","DOIUrl":"10.1177/00131644251399588","url":null,"abstract":"<p><p>Researchers in the behavioral, educational, and social sciences often aim to analyze relationships among latent variables. Structural equation modeling (SEM) is widely regarded as the gold standard for this purpose. A straightforward alternative for estimating the structural model parameters is uncorrected factor score regression (UFSR), where factor scores are first computed and then employed in regression or path analysis. Unfortunately, the most commonly used factor scores (i.e., Regression and Bartlett factor scores) may yield biased estimates and invalid inferences when using this approach. In recent years, factor score regression (FSR) has enjoyed several methodological advancements to address this inconsistency. Despite these advancements, the use of FSR with correlation-preserving factor scores, here termed consistent factor score regression (cFSR), has received limited attention. In this paper, we revisit cFSR and compare its advantages and disadvantages relative to other recent FSR and SEM methods. We conducted an extensive simulation study comparing cFSR with other estimation approaches, assessing their performance in terms of convergence rate, bias, efficiency, and type I error rate. The findings indicate that cFSR outperforms UFSR while maintaining the conceptual simplicity of UFSR. We encourage behavioral, educational, and social science researchers to avoid UFSR and adopt cFSR as an alternative to SEM.</p>","PeriodicalId":11502,"journal":{"name":"Educational and Psychological Measurement","volume":" ","pages":"00131644251399588"},"PeriodicalIF":2.3,"publicationDate":"2026-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12774818/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145932674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Educational and Psychological Measurement
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1