Pub Date : 2025-08-01Epub Date: 2023-08-10DOI: 10.1037/met0000591
James L Peugh, Kaylee Litson, David F Feldon
Decades of published methodological research have shown the chi-square test of model fit performs inconsistently and unreliably as a determinant of structural equation model (SEM) fit. Likewise, SEM indices of model fit, such as comparative fit index (CFI) and root-mean-square error of approximation (RMSEA) also perform inconsistently and unreliably. Despite rather unreliable ways to statistically assess model fit, researchers commonly rely on these methods for lack of a suitable inferential alternative. Marcoulides and Yuan (2017) have proposed the first inferential test of SEM fit in many years: an equivalence test adaptation of the RMSEA and CFI indices (i.e., RMSEAt and CFIt). However, the ability of this equivalence testing approach to accurately judge acceptable and unacceptable model fit has not been empirically tested. This fully crossed Monte Carlo simulation evaluated the accuracy of equivalence testing combining many of the same independent variable (IV) conditions used in previous fit index simulation studies, including sample size (N = 100-1,000), model specification (correctly specified or misspecified), model type (confirmatory factor analysis [CFA], path analysis, or SEM), number of variables analyzed (low or high), data distribution (normal or skewed), and missing data (none, 10%, or 25%). Results show equivalence testing performs rather inconsistently and unreliably across IV conditions, with acceptable or unacceptable RMSEAt and CFIt model fit index values often being contingent on complex interactions among conditions. Proportional z-tests and logistic regression analyses indicated that equivalence tests of model fit are problematic under multiple conditions, especially those where models are mildly misspecified. Recommendations for researchers are offered, but with the provision that they be used with caution until more research and development is available. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
几十年发表的方法学研究表明,模型拟合的卡方检验作为结构方程模型(SEM)拟合的决定因素表现得不一致和不可靠。同样,模型拟合的SEM指标,如比较拟合指数(CFI)和近似均方根误差(RMSEA)也表现不一致和不可靠。尽管统计评估模型拟合的方法相当不可靠,但由于缺乏合适的推断替代方法,研究人员通常依赖这些方法。Marcoulides和Yuan(2017)提出了多年来第一个SEM拟合的推理检验:RMSEA和CFI指数(即RMSEAt和CFIt)的等效检验。然而,这种等效检验方法准确判断可接受和不可接受模型拟合的能力尚未得到实证检验。这个完全交叉的蒙特卡罗模拟评估了等效检验的准确性,结合了许多在以前的拟合指数模拟研究中使用的相同的自变量(IV)条件,包括样本量(N = 100- 1000)、模型规格(正确指定或错误指定)、模型类型(验证性因子分析[CFA]、路径分析或SEM)、分析的变量数量(低或高)、数据分布(正态或偏态)和缺失数据(无、10%或25%)。结果表明,等效性测试在IV条件下执行得相当不一致和不可靠,RMSEAt和CFIt模型拟合指数值通常取决于条件之间的复杂相互作用。比例z检验和逻辑回归分析表明,在多种条件下,模型拟合的等效检验存在问题,特别是在模型轻度错误指定的情况下。为研究人员提供了建议,但规定在有更多的研究和发展可用之前,要谨慎使用这些建议。(PsycInfo Database Record (c) 2025 APA,版权所有)。
{"title":"Equivalence testing to judge model fit: A Monte Carlo simulation.","authors":"James L Peugh, Kaylee Litson, David F Feldon","doi":"10.1037/met0000591","DOIUrl":"10.1037/met0000591","url":null,"abstract":"<p><p>Decades of published methodological research have shown the chi-square test of model fit performs inconsistently and unreliably as a determinant of structural equation model (SEM) fit. Likewise, SEM indices of model fit, such as comparative fit index (CFI) and root-mean-square error of approximation (RMSEA) also perform inconsistently and unreliably. Despite rather unreliable ways to statistically assess model fit, researchers commonly rely on these methods for lack of a suitable inferential alternative. Marcoulides and Yuan (2017) have proposed the first inferential test of SEM fit in many years: an equivalence test adaptation of the RMSEA and CFI indices (i.e., RMSEA<sub><i>t</i></sub> and CFI<i><sub>t</sub></i>). However, the ability of this equivalence testing approach to accurately judge acceptable and unacceptable model fit has not been empirically tested. This fully crossed Monte Carlo simulation evaluated the accuracy of equivalence testing combining many of the same independent variable (IV) conditions used in previous fit index simulation studies, including sample size (<i>N</i> = 100-1,000), model specification (correctly specified or misspecified), model type (confirmatory factor analysis [CFA], path analysis, or SEM), number of variables analyzed (low or high), data distribution (normal or skewed), and missing data (none, 10%, or 25%). Results show equivalence testing performs rather inconsistently and unreliably across IV conditions, with acceptable or unacceptable RMSEA<i><sub>t</sub></i> and CFIt model fit index values often being contingent on complex interactions among conditions. Proportional <i>z</i>-tests and logistic regression analyses indicated that equivalence tests of model fit are problematic under multiple conditions, especially those where models are mildly misspecified. Recommendations for researchers are offered, but with the provision that they be used with caution until more research and development is available. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"888-925"},"PeriodicalIF":7.8,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10339181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2023-07-20DOI: 10.1037/met0000581
Antoinette D A Kroes, Jason R Finley
Omega squared (ω^2) is a measure of effect size for analysis of variance (ANOVA) designs. It is less biased than eta squared, but reported less often. This is in part due to lack of clear guidance on how to calculate it. In this paper, we discuss the logic behind effect size measures, the problem with eta squared, the history of omega squared, and why it has been underused. We then provide a user-friendly guide to omega squared and partial omega squared for ANOVA designs with fixed factors, including one-way, two-way, and three-way designs, using within-subjects factors and/or between-subjects factors. We show how to calculate omega squared using output from SPSS. We provide information on the calculation of confidence intervals. We examine the problems of nonadditivity, and intrinsic versus extrinsic factors. We argue that statistical package developers could play an important role in making the calculation of omega squared easier. Finally, we recommend that researchers report the formulas used in calculating effect sizes, include confidence intervals if possible, and include ANOVA tables in the online supplemental materials of their work. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
平方(ω^2)是方差分析(ANOVA)设计的效应大小的度量。它的偏差小于平方,但报告的频率较低。这在一定程度上是由于缺乏关于如何计算的明确指导。在本文中,我们讨论了效应大小测量背后的逻辑,平方的问题,平方的历史,以及为什么它没有得到充分利用。然后,我们为具有固定因素的方差分析设计提供了一个用户友好的omega平方和部分omega平方指南,包括单向,双向和三向设计,使用受试者内因素和/或受试者之间因素。我们展示了如何使用SPSS的输出来计算omega的平方。我们提供了计算置信区间的信息。我们研究了不可加性问题,以及内在因素与外在因素的对比。我们认为统计软件包开发人员可以在简化计算平方方面发挥重要作用。最后,我们建议研究人员报告用于计算效应量的公式,如果可能的话,包括置信区间,并在他们的工作的在线补充材料中包括方差分析表。(PsycInfo Database Record (c) 2025 APA,版权所有)。
{"title":"Demystifying omega squared: Practical guidance for effect size in common analysis of variance designs.","authors":"Antoinette D A Kroes, Jason R Finley","doi":"10.1037/met0000581","DOIUrl":"10.1037/met0000581","url":null,"abstract":"<p><p>Omega squared (ω^2) is a measure of effect size for analysis of variance (ANOVA) designs. It is less biased than eta squared, but reported less often. This is in part due to lack of clear guidance on how to calculate it. In this paper, we discuss the logic behind effect size measures, the problem with eta squared, the history of omega squared, and why it has been underused. We then provide a user-friendly guide to omega squared and partial omega squared for ANOVA designs with fixed factors, including one-way, two-way, and three-way designs, using within-subjects factors and/or between-subjects factors. We show how to calculate omega squared using output from SPSS. We provide information on the calculation of confidence intervals. We examine the problems of nonadditivity, and intrinsic versus extrinsic factors. We argue that statistical package developers could play an important role in making the calculation of omega squared easier. Finally, we recommend that researchers report the formulas used in calculating effect sizes, include confidence intervals if possible, and include ANOVA tables in the online supplemental materials of their work. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"866-887"},"PeriodicalIF":7.8,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9840882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2023-08-10DOI: 10.1037/met0000603
Lydia Craig Aulisi, Hannah M Markell-Goldstein, Jose M Cortina, Carol M Wong, Xue Lei, Cyrus K Foroughi
Meta-analyses in the psychological sciences typically examine moderators that may explain heterogeneity in effect sizes. One of the most commonly examined moderators is gender. Overall, tests of gender as a moderator are rarely significant, which may be because effects rarely differ substantially between men and women. While this may be true in some cases, we also suggest that the lack of significant findings may be attributable to the way in which gender is examined as a meta-analytic moderator, such that detecting moderating effects is very unlikely even when such effects are substantial in magnitude. More specifically, we suggest that lack of between-primary study variance in gender composition makes it exceedingly difficult to detect moderation. That is, because primary studies tend to have similar male-to-female ratios, there is very little variance in gender composition between primaries, making it nearly impossible to detect between-study differences in the relationship of interest as a function of gender. In the present article, we report results from two studies: (a) a meta-meta-analysis in which we demonstrate the magnitude of this problem by computing the between-study variance in gender composition across 286 meta-analytic moderation tests from 50 meta-analyses, and (b) a Monte Carlo simulation study in which we show that this lack of variance results in near-zero moderator effects even when male-female differences in correlations are quite large. Our simulations are also used to show the value of single-gender studies for detecting moderating effects. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
心理科学的荟萃分析通常检查可能解释效应大小异质性的调节因子。最常被检查的调节因素之一是性别。总的来说,性别作为调节因素的测试很少有意义,这可能是因为男性和女性之间的影响很少有实质性差异。虽然这在某些情况下可能是正确的,但我们也认为,缺乏重大发现可能归因于性别作为元分析调节因素的检验方式,这样即使在这种影响很大的情况下,也不太可能检测到调节作用。更具体地说,我们认为缺乏性别构成的主要研究之间的差异使得很难检测到适度。也就是说,由于初级研究往往有相似的男女比例,初级研究之间的性别构成差异很小,因此几乎不可能检测到兴趣关系作为性别函数的研究之间的差异。在本文中,我们报告了两项研究的结果:(a)一项荟萃分析,我们通过计算来自50项荟萃分析的286项荟萃分析调节测试的性别组成的研究间方差来证明这个问题的严重性;(b)一项蒙特卡罗模拟研究,我们表明,即使在男女相关性差异相当大的情况下,这种方差的缺乏也会导致接近零的调节效应。我们的模拟也被用来显示单性别研究在检测调节效应方面的价值。(PsycInfo Database Record (c) 2025 APA,版权所有)。
{"title":"Detecting gender as a moderator in meta-analysis: The problem of restricted between-study variance.","authors":"Lydia Craig Aulisi, Hannah M Markell-Goldstein, Jose M Cortina, Carol M Wong, Xue Lei, Cyrus K Foroughi","doi":"10.1037/met0000603","DOIUrl":"10.1037/met0000603","url":null,"abstract":"<p><p>Meta-analyses in the psychological sciences typically examine moderators that may explain heterogeneity in effect sizes. One of the most commonly examined moderators is gender. Overall, tests of gender as a moderator are rarely significant, which may be because effects rarely differ substantially between men and women. While this may be true in some cases, we also suggest that the lack of significant findings may be attributable to the way in which gender is examined as a meta-analytic moderator, such that detecting moderating effects is very unlikely even when such effects are substantial in magnitude. More specifically, we suggest that lack of between-primary study variance in gender composition makes it exceedingly difficult to detect moderation. That is, because primary studies tend to have similar male-to-female ratios, there is very little variance in gender composition between primaries, making it nearly impossible to detect between-study differences in the relationship of interest as a function of gender. In the present article, we report results from two studies: (a) a meta-meta-analysis in which we demonstrate the magnitude of this problem by computing the between-study variance in gender composition across 286 meta-analytic moderation tests from 50 meta-analyses, and (b) a Monte Carlo simulation study in which we show that this lack of variance results in near-zero moderator effects even when male-female differences in correlations are quite large. Our simulations are also used to show the value of single-gender studies for detecting moderating effects. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"687-719"},"PeriodicalIF":7.8,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9967420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-01Epub Date: 2023-07-27DOI: 10.1037/met0000597
Peter M Steiner, Patrick Sheehan, Vivian C Wong
Given recent evidence challenging the replicability of results in the social and behavioral sciences, critical questions have been raised about appropriate measures for determining replication success in comparing effect estimates across studies. At issue is the fact that conclusions about replication success often depend on the measure used for evaluating correspondence in results. Despite the importance of choosing an appropriate measure, there is still no widespread agreement about which measures should be used. This article addresses these questions by describing formally the most commonly used measures for assessing replication success, and by comparing their performance in different contexts according to their replication probabilities-that is, the probability of obtaining replication success given study-specific settings. The measures may be characterized broadly as conclusion-based approaches, which assess the congruence of two independent studies' conclusions about the presence of an effect, and distance-based approaches, which test for a significant difference or equivalence of two effect estimates. We also introduce a new measure for assessing replication success called the correspondence test, which combines a difference and equivalence test in the same framework. To help researchers plan prospective replication efforts, we provide closed formulas for power calculations that can be used to determine the minimum detectable effect size (and thus, sample sizes) for each study so that a predetermined minimum replication probability can be achieved. Finally, we use a replication data set from the Open Science Collaboration (2015) to demonstrate the extent to which conclusions about replication success depend on the correspondence measure selected. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
鉴于最近的证据对社会和行为科学结果的可复制性提出了挑战,在比较不同研究的效果估计时,确定复制成功的适当措施提出了关键问题。争论的焦点在于,关于复制成功与否的结论往往取决于用于评估结果一致性的测量方法。尽管选择适当的措施很重要,但对于应该使用哪些措施仍然没有广泛的共识。本文通过正式描述用于评估复制成功的最常用度量,并根据它们的复制概率(即给定特定研究设置的获得复制成功的概率)比较它们在不同上下文中的性能,来解决这些问题。这些措施可以被广泛地描述为基于结论的方法,评估两个独立研究关于效应存在的结论的一致性,以及基于距离的方法,测试两个效应估计的显着差异或等效性。我们还引入了一种评估复制成功的新方法,称为对应测试,它在同一框架中结合了差异测试和等效测试。为了帮助研究人员计划前瞻性的复制工作,我们提供了功率计算的封闭公式,可用于确定每个研究的最小可检测效应大小(从而确定样本量),从而可以实现预定的最小复制概率。最后,我们使用开放科学协作(2015)的复制数据集来证明关于复制成功的结论在多大程度上取决于所选择的对应度量。(PsycInfo Database Record (c) 2025 APA,版权所有)。
{"title":"Correspondence measures for assessing replication success.","authors":"Peter M Steiner, Patrick Sheehan, Vivian C Wong","doi":"10.1037/met0000597","DOIUrl":"10.1037/met0000597","url":null,"abstract":"<p><p>Given recent evidence challenging the replicability of results in the social and behavioral sciences, critical questions have been raised about appropriate measures for determining replication success in comparing effect estimates across studies. At issue is the fact that conclusions about replication success often depend on the measure used for evaluating correspondence in results. Despite the importance of choosing an appropriate measure, there is still no widespread agreement about which measures should be used. This article addresses these questions by describing formally the most commonly used measures for assessing replication success, and by comparing their performance in different contexts according to their replication probabilities-that is, the probability of obtaining replication success given study-specific settings. The measures may be characterized broadly as conclusion-based approaches, which assess the congruence of two independent studies' conclusions about the presence of an effect, and distance-based approaches, which test for a significant difference or equivalence of two effect estimates. We also introduce a new measure for assessing replication success called the correspondence test, which combines a difference and equivalence test in the same framework. To help researchers plan prospective replication efforts, we provide closed formulas for power calculations that can be used to determine the minimum detectable effect size (and thus, sample sizes) for each study so that a predetermined minimum replication probability can be achieved. Finally, we use a replication data set from the Open Science Collaboration (2015) to demonstrate the extent to which conclusions about replication success depend on the correspondence measure selected. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"793-814"},"PeriodicalIF":7.8,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10259359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yutian T. Thompson, Yaqi Li, Hairong Song, David Bard
{"title":"Joint variable selection in generalized linear mixed models with random regularized penalized quasi-likelihood technique.","authors":"Yutian T. Thompson, Yaqi Li, Hairong Song, David Bard","doi":"10.1037/met0000783","DOIUrl":"https://doi.org/10.1037/met0000783","url":null,"abstract":"","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":"52 1","pages":""},"PeriodicalIF":7.0,"publicationDate":"2025-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144748224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Supplemental Material for An Explainable Artificial Intelligence Handbook for Psychologists: Methods, Opportunities, and Challenges","authors":"","doi":"10.1037/met0000772.supp","DOIUrl":"https://doi.org/10.1037/met0000772.supp","url":null,"abstract":"","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":"27 1","pages":""},"PeriodicalIF":7.0,"publicationDate":"2025-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144715713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Supplemental Material for Joint Variable Selection in Generalized Linear Mixed Models With Random Regularized Penalized Quasi-Likelihood Technique","authors":"","doi":"10.1037/met0000783.supp","DOIUrl":"https://doi.org/10.1037/met0000783.supp","url":null,"abstract":"","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":"283 1","pages":""},"PeriodicalIF":7.0,"publicationDate":"2025-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144715438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This article proposes an approach for detecting multivariate outliers that combines robust estimation methods with signed detection information. Our method uses the Mahalanobis distance to quantify each observation's extremeness from the expected value relative to the covariance matrix, and we leverage robust estimation tools, i.e., the minimum covariance determinant, to estimate the mean vector and covariance matrix used in the Mahalanobis distance calculation. Furthermore, we incorporate a signing element into the distance calculation to give researchers greater control over the specific regions of multivariate space that should be prioritized when searching for outliers, which allows for more targeted risk assessment and classification. Lastly, we unify the robust and signed elements into a framework that can be used within bilinear models such as principal components analysis and factor analysis. Using simulated and real data examples, we demonstrate that the proposed approach can result in improved risk assessment and outlier detection, particularly when the sample is contaminated with a moderate-to-large number of outliers that have noteworthy contamination strengths. Overall, our results show that making use of a robust method when assessing multivariate risk leads to more accurate estimates, particularly when combined with relevant signing information. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
本文提出了一种检测多变量异常值的方法,该方法结合了鲁棒估计方法和签名检测信息。我们的方法使用马氏距离从相对于协方差矩阵的期望值来量化每个观测值的极值,并且我们利用鲁棒估计工具,即最小协方差行列式,来估计马氏距离计算中使用的平均向量和协方差矩阵。此外,我们在距离计算中加入了一个签名元素,使研究人员能够更好地控制在搜索异常值时应该优先考虑的多元空间的特定区域,从而允许更有针对性的风险评估和分类。最后,我们将鲁棒元素和签名元素统一到一个框架中,该框架可用于双线性模型,如主成分分析和因子分析。通过模拟和真实数据示例,我们证明了所提出的方法可以改进风险评估和异常值检测,特别是当样本被具有显著污染强度的中等到大量异常值污染时。总体而言,我们的研究结果表明,在评估多变量风险时使用稳健的方法可以获得更准确的估计,特别是在与相关签名信息相结合时。(PsycInfo Database Record (c) 2025 APA,版权所有)。
{"title":"Robust detection of signed outliers in multivariate data with applications to early identification of risk for autism.","authors":"Jesus E Delgado,Jed T Elison,Nathaniel E Helwig","doi":"10.1037/met0000775","DOIUrl":"https://doi.org/10.1037/met0000775","url":null,"abstract":"This article proposes an approach for detecting multivariate outliers that combines robust estimation methods with signed detection information. Our method uses the Mahalanobis distance to quantify each observation's extremeness from the expected value relative to the covariance matrix, and we leverage robust estimation tools, i.e., the minimum covariance determinant, to estimate the mean vector and covariance matrix used in the Mahalanobis distance calculation. Furthermore, we incorporate a signing element into the distance calculation to give researchers greater control over the specific regions of multivariate space that should be prioritized when searching for outliers, which allows for more targeted risk assessment and classification. Lastly, we unify the robust and signed elements into a framework that can be used within bilinear models such as principal components analysis and factor analysis. Using simulated and real data examples, we demonstrate that the proposed approach can result in improved risk assessment and outlier detection, particularly when the sample is contaminated with a moderate-to-large number of outliers that have noteworthy contamination strengths. Overall, our results show that making use of a robust method when assessing multivariate risk leads to more accurate estimates, particularly when combined with relevant signing information. (PsycInfo Database Record (c) 2025 APA, all rights reserved).","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":"26 1","pages":""},"PeriodicalIF":7.0,"publicationDate":"2025-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144669552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Adapting methods for correcting selective reporting bias in meta-analysis of dependent effect sizes.","authors":"Man Chen, James E. Pustejovsky","doi":"10.1037/met0000773","DOIUrl":"https://doi.org/10.1037/met0000773","url":null,"abstract":"","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":"107 1","pages":""},"PeriodicalIF":7.0,"publicationDate":"2025-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144602945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}