Björn S Siepe, František Bartoš, Tim P Morris, Anne-Laure Boulesteix, Daniel W Heck, Samuel Pawel
Simulation studies are widely used for evaluating the performance of statistical methods in psychology. However, the quality of simulation studies can vary widely in terms of their design, execution, and reporting. In order to assess the quality of typical simulation studies in psychology, we reviewed 321 articles published in Psychological Methods, Behavior Research Methods, and Multivariate Behavioral Research in 2021 and 2022, among which 100/321 = 31.2% report a simulation study. We find that many articles do not provide complete and transparent information about key aspects of the study, such as justifications for the number of simulation repetitions, Monte Carlo uncertainty estimates, or code and data to reproduce the simulation studies. To address this problem, we provide a summary of the ADEMP (aims, data-generating mechanism, estimands and other targets, methods, performance measures) design and reporting framework from Morris et al. (2019) adapted to simulation studies in psychology. Based on this framework, we provide ADEMP-PreReg, a step-by-step template for researchers to use when designing, potentially preregistering, and reporting their simulation studies. We give formulae for estimating common performance measures, their Monte Carlo standard errors, and for calculating the number of simulation repetitions to achieve a desired Monte Carlo standard error. Finally, we give a detailed tutorial on how to apply the ADEMP framework in practice using an example simulation study on the evaluation of methods for the analysis of pre-post measurement experiments. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
模拟研究被广泛用于评估心理学统计方法的性能。然而,模拟研究在设计、执行和报告方面的质量可能存在很大差异。为了评估心理学中典型模拟研究的质量,我们查阅了 2021 年和 2022 年发表在《心理学方法》、《行为研究方法》和《多元行为研究》上的 321 篇文章,其中 100/321 = 31.2% 的文章报告了模拟研究。我们发现,许多文章没有提供完整、透明的研究关键方面的信息,如模拟重复次数的理由、蒙特卡罗不确定性估计或重现模拟研究的代码和数据。为了解决这个问题,我们总结了 Morris 等人(2019 年)的 ADEMP(目的、数据生成机制、估计值和其他目标、方法、绩效衡量)设计和报告框架,并将其调整为心理学中的模拟研究。在此框架基础上,我们提供了 ADEMP-PreReg,一个供研究人员在设计、预注册和报告模拟研究时使用的分步模板。我们给出了估算常见性能指标、其蒙特卡洛标准误差以及计算达到所需的蒙特卡洛标准误差所需的模拟重复次数的公式。最后,我们通过一个关于事后测量实验分析方法评估的模拟研究实例,详细介绍了如何在实践中应用 ADEMP 框架。(PsycInfo Database Record (c) 2024 APA, all rights reserved)。
{"title":"Simulation studies for methodological research in psychology: A standardized template for planning, preregistration, and reporting.","authors":"Björn S Siepe, František Bartoš, Tim P Morris, Anne-Laure Boulesteix, Daniel W Heck, Samuel Pawel","doi":"10.1037/met0000695","DOIUrl":"https://doi.org/10.1037/met0000695","url":null,"abstract":"<p><p>Simulation studies are widely used for evaluating the performance of statistical methods in psychology. However, the quality of simulation studies can vary widely in terms of their design, execution, and reporting. In order to assess the quality of typical simulation studies in psychology, we reviewed 321 articles published in <i>Psychological Methods, Behavior Research Methods, and Multivariate Behavioral Research</i> in 2021 and 2022, among which 100/321 = 31.2% report a simulation study. We find that many articles do not provide complete and transparent information about key aspects of the study, such as justifications for the number of simulation repetitions, Monte Carlo uncertainty estimates, or code and data to reproduce the simulation studies. To address this problem, we provide a summary of the ADEMP (aims, data-generating mechanism, estimands and other targets, methods, performance measures) design and reporting framework from Morris et al. (2019) adapted to simulation studies in psychology. Based on this framework, we provide ADEMP-PreReg, a step-by-step template for researchers to use when designing, potentially preregistering, and reporting their simulation studies. We give formulae for estimating common performance measures, their Monte Carlo standard errors, and for calculating the number of simulation repetitions to achieve a desired Monte Carlo standard error. Finally, we give a detailed tutorial on how to apply the ADEMP framework in practice using an example simulation study on the evaluation of methods for the analysis of pre-post measurement experiments. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":""},"PeriodicalIF":7.6,"publicationDate":"2024-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142626859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2023-04-27DOI: 10.1037/met0000564
Wen Wei Loh, Dongning Ren
Valid inference of cause-and-effect relations in observational studies necessitates adjusting for common causes of the focal predictor (i.e., treatment) and the outcome. When such common causes, henceforth termed confounders, remain unadjusted for, they generate spurious correlations that lead to biased causal effect estimates. But routine adjustment for all available covariates, when only a subset are truly confounders, is known to yield potentially inefficient and unstable estimators. In this article, we introduce a data-driven confounder selection strategy that focuses on stable estimation of the treatment effect. The approach exploits the causal knowledge that after adjusting for confounders to eliminate all confounding biases, adding any remaining non-confounding covariates associated with only treatment or outcome, but not both, should not systematically change the effect estimator. The strategy proceeds in two steps. First, we prioritize covariates for adjustment by probing how strongly each covariate is associated with treatment and outcome. Next, we gauge the stability of the effect estimator by evaluating its trajectory adjusting for different covariate subsets. The smallest subset that yields a stable effect estimate is then selected. Thus, the strategy offers direct insight into the (in)sensitivity of the effect estimator to the chosen covariates for adjustment. The ability to correctly select confounders and yield valid causal inferences following data-driven covariate selection is evaluated empirically using extensive simulation studies. Furthermore, we compare the introduced method empirically with routine variable selection methods. Finally, we demonstrate the procedure using two publicly available real-world datasets. A step-by-step practical guide with user-friendly R functions is included. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
要在观察性研究中有效推断因果关系,就必须对重点预测因子(即治疗)和结果的共同原因进行调整。如果不对这些共同原因(以下称为混杂因素)进行调整,就会产生虚假的相关性,导致因果效应估计值出现偏差。但众所周知,当只有一部分是真正的混杂因素时,对所有可用的协变量进行常规调整可能会产生低效和不稳定的估计值。在本文中,我们介绍了一种数据驱动的混杂因素选择策略,其重点是稳定估计治疗效果。该方法利用的因果知识是,在调整混杂因素以消除所有混杂偏差后,添加任何仅与治疗或结果相关而非两者相关的非混杂协变量,都不应系统性地改变效果估计值。该策略分两步进行。首先,我们通过探究每个协变量与治疗和结果的关联程度,确定需要调整的协变量的优先级。接下来,我们通过评估不同协变量子集的调整轨迹来衡量效应估计值的稳定性。然后选择能产生稳定效应估计值的最小子集。因此,该策略可以直接洞察效应估计值对所选协变因素调整的(不)敏感性。通过大量的模拟研究,我们对数据驱动协变量选择后正确选择混杂因素并得出有效因果推论的能力进行了实证评估。此外,我们还将引入的方法与常规变量选择方法进行了实证比较。最后,我们使用两个公开的真实数据集演示了这一过程。此外,我们还提供了使用方便的 R 函数的分步实践指南。(PsycInfo Database Record (c) 2024 APA, all rights reserved)。
{"title":"Data-driven covariate selection for confounding adjustment by focusing on the stability of the effect estimator.","authors":"Wen Wei Loh, Dongning Ren","doi":"10.1037/met0000564","DOIUrl":"10.1037/met0000564","url":null,"abstract":"<p><p>Valid inference of cause-and-effect relations in observational studies necessitates adjusting for common causes of the focal predictor (i.e., treatment) and the outcome. When such common causes, henceforth termed confounders, remain unadjusted for, they generate spurious correlations that lead to biased causal effect estimates. But routine adjustment for all available covariates, when only a subset are truly confounders, is known to yield potentially inefficient and unstable estimators. In this article, we introduce a data-driven confounder selection strategy that focuses on stable estimation of the treatment effect. The approach exploits the causal knowledge that after adjusting for confounders to eliminate all confounding biases, adding any remaining non-confounding covariates associated with only treatment or outcome, but not both, should not systematically change the effect estimator. The strategy proceeds in two steps. First, we prioritize covariates for adjustment by probing how strongly each covariate is associated with treatment and outcome. Next, we gauge the stability of the effect estimator by evaluating its trajectory adjusting for different covariate subsets. The smallest subset that yields a stable effect estimate is then selected. Thus, the strategy offers direct insight into the (in)sensitivity of the effect estimator to the chosen covariates for adjustment. The ability to correctly select confounders and yield valid causal inferences following data-driven covariate selection is evaluated empirically using extensive simulation studies. Furthermore, we compare the introduced method empirically with routine variable selection methods. Finally, we demonstrate the procedure using two publicly available real-world datasets. A step-by-step practical guide with user-friendly R functions is included. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"947-966"},"PeriodicalIF":7.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9356535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2022-10-13DOI: 10.1037/met0000534
David Jendryczko, Fridtjof W Nussbeck
The present contribution provides a tutorial for the estimation of the social relations model (SRM) by means of structural equation modeling (SEM). In the overarching SEM-framework, the SRM without roles (with interchangeable dyads) is derived as a more restrictive form of the SRM with roles (with noninterchangeable dyads). Starting with the simplest type of the SRM for one latent construct assessed by one manifest round-robin indicator, we show how the model can be extended to multiple constructs each measured by multiple indicators. We illustrate a multiple constructs multiple indicators SEM SRM both with and without roles with simulated data and explain the parameter interpretations. We present how testing the substantial model assumptions can be disentangled from testing the interchangeability of dyads. Additionally, we point out modeling strategies that adhere to cases in which only some members of a group can be differentiated with regards to their roles (i.e., only some group members are noninterchangeable). In the online supplemental materials, we provide concrete examples of specific modeling problems and their implementation into statistical software (Mplus, lavaan, and OpenMx). Advantages, caveats, possible extensions, and limitations in comparison with alternative modeling options are discussed. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
本论文提供了通过结构方程建模(SEM)估算社会关系模型(SRM)的教程。在 SEM 的总体框架中,无角色 SRM(可互换的二人组)是有角色 SRM(不可互换的二人组)的一种限制性更强的形式。我们从最简单的 SRM 模型(由一个显性循环指标评估一个潜在构造)开始,展示了如何将该模型扩展到由多个指标测量的多个构造。我们用模拟数据说明了有角色和无角色的多构式多指标 SEM SRM,并解释了参数解释。我们介绍了如何将对实质性模型假设的测试与对二元组互换性的测试区分开来。此外,我们还指出了一些建模策略,这些策略适用于小组中只有部分成员的角色可以区分的情况(即只有部分小组成员不可互换)。在在线补充材料中,我们提供了具体建模问题及其在统计软件(Mplus、lavaan 和 OpenMx)中实现的具体示例。我们还讨论了与其他建模方案相比的优势、注意事项、可能的扩展和局限性。(PsycInfo Database Record (c) 2024 APA, 版权所有)。
{"title":"Estimating and investigating multiple constructs multiple indicators social relations models with and without roles within the traditional structural equation modeling framework: A tutorial.","authors":"David Jendryczko, Fridtjof W Nussbeck","doi":"10.1037/met0000534","DOIUrl":"10.1037/met0000534","url":null,"abstract":"<p><p>The present contribution provides a tutorial for the estimation of the social relations model (SRM) by means of structural equation modeling (SEM). In the overarching SEM-framework, the SRM without roles (with interchangeable dyads) is derived as a more restrictive form of the SRM with roles (with noninterchangeable dyads). Starting with the simplest type of the SRM for one latent construct assessed by one manifest round-robin indicator, we show how the model can be extended to multiple constructs each measured by multiple indicators. We illustrate a multiple constructs multiple indicators SEM SRM both with and without roles with simulated data and explain the parameter interpretations. We present how testing the substantial model assumptions can be disentangled from testing the interchangeability of dyads. Additionally, we point out modeling strategies that adhere to cases in which only some members of a group can be differentiated with regards to their roles (i.e., only some group members are noninterchangeable). In the online supplemental materials, we provide concrete examples of specific modeling problems and their implementation into statistical software (Mplus, lavaan, and OpenMx). Advantages, caveats, possible extensions, and limitations in comparison with alternative modeling options are discussed. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"919-946"},"PeriodicalIF":7.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9371931","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2022-09-01DOI: 10.1037/met0000516
Debby Ten Hove, Terrence D Jorgensen, L Andries van der Ark
Several intraclass correlation coefficients (ICCs) are available to assess the interrater reliability (IRR) of observational measurements. Selecting an ICC is complicated, and existing guidelines have three major limitations. First, they do not discuss incomplete designs, in which raters partially vary across subjects. Second, they provide no coherent perspective on the error variance in an ICC, clouding the choice between the available coefficients. Third, the distinction between fixed or random raters is often misunderstood. Based on generalizability theory (GT), we provide updated guidelines on selecting an ICC for IRR, which are applicable to both complete and incomplete observational designs. We challenge conventional wisdom about ICCs for IRR by claiming that raters should seldom (if ever) be considered fixed. Also, we clarify how to interpret ICCs in the case of unbalanced and incomplete designs. We explain four choices a researcher needs to make when selecting an ICC for IRR, and guide researchers through these choices by means of a flowchart, which we apply to three empirical examples from clinical and developmental domains. In the Discussion, we provide guidance in reporting, interpreting, and estimating ICCs, and propose future directions for research into the ICCs for IRR. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"Updated guidelines on selecting an intraclass correlation coefficient for interrater reliability, with applications to incomplete observational designs.","authors":"Debby Ten Hove, Terrence D Jorgensen, L Andries van der Ark","doi":"10.1037/met0000516","DOIUrl":"10.1037/met0000516","url":null,"abstract":"<p><p>Several intraclass correlation coefficients (ICCs) are available to assess the interrater reliability (IRR) of observational measurements. Selecting an ICC is complicated, and existing guidelines have three major limitations. First, they do not discuss incomplete designs, in which raters partially vary across subjects. Second, they provide no coherent perspective on the error variance in an ICC, clouding the choice between the available coefficients. Third, the distinction between fixed or random raters is often misunderstood. Based on generalizability theory (GT), we provide updated guidelines on selecting an ICC for IRR, which are applicable to both complete and incomplete observational designs. We challenge conventional wisdom about ICCs for IRR by claiming that raters should seldom (if ever) be considered fixed. Also, we clarify how to interpret ICCs in the case of unbalanced and incomplete designs. We explain four choices a researcher needs to make when selecting an ICC for IRR, and guide researchers through these choices by means of a flowchart, which we apply to three empirical examples from clinical and developmental domains. In the Discussion, we provide guidance in reporting, interpreting, and estimating ICCs, and propose future directions for research into the ICCs for IRR. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"967-979"},"PeriodicalIF":7.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9290331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2022-10-06DOI: 10.1037/met0000530
Kenneth A Bollen, Adam G Lilly, Lan Luo
It is common practice for psychologists to specify models with latent variables to represent concepts that are difficult to directly measure. Each latent variable needs a scale, and the most popular method of scaling as well as the default in most structural equation modeling (SEM) software uses a scaling or reference indicator. Much of the time, the choice of which indicator to use for this purpose receives little attention, and many analysts use the first indicator without considering whether there are better choices. When all indicators of the latent variable have essentially the same properties, then the choice matters less. But when this is not true, we could benefit from scaling indicator guidelines. Our article first demonstrates why latent variables need a scale. We then propose a set of criteria and accompanying diagnostic tools that can assist researchers in making informed decisions about scaling indicators. The criteria for a good scaling indicator include high face validity, high correlation with the latent variable, factor complexity of one, no correlated errors, no direct effects with other indicators, a minimal number of significant overidentification equation tests and modification indices, and invariance across groups and time. We demonstrate these criteria and diagnostics using two empirical examples and provide guidance on navigating conflicting results among criteria. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
心理学家通常会用潜变量来指定模型,以表示难以直接测量的概念。每个潜变量都需要一个标度,而最常用的标度方法以及大多数结构方程建模(SEM)软件的默认值都使用标度或参考指标。在很多情况下,选择使用哪个指标并没有引起足够的重视,很多分析师使用第一个指标,而不考虑是否有更好的选择。当潜在变量的所有指标都具有基本相同的属性时,选择就不那么重要了。但当情况并非如此时,我们就可以从缩放指标指南中获益。我们的文章首先说明了为什么潜变量需要标度。然后,我们提出了一套标准和相应的诊断工具,可以帮助研究人员就标度指标做出明智的决定。好的标度指标的标准包括:高表面效度、与潜变量高度相关、因子复杂度为一、无相关误差、与其他指标无直接影响、最小数量的显著过度识别方程测试和修正指数,以及跨组和跨时间的不变性。我们通过两个实证案例来展示这些标准和诊断方法,并为如何在标准之间找到相互矛盾的结果提供指导。(PsycInfo Database Record (c) 2022 APA,保留所有权利)。
{"title":"Selecting scaling indicators in structural equation models (sems).","authors":"Kenneth A Bollen, Adam G Lilly, Lan Luo","doi":"10.1037/met0000530","DOIUrl":"10.1037/met0000530","url":null,"abstract":"<p><p>It is common practice for psychologists to specify models with latent variables to represent concepts that are difficult to directly measure. Each latent variable needs a scale, and the most popular method of scaling as well as the default in most structural equation modeling (SEM) software uses a scaling or reference indicator. Much of the time, the choice of which indicator to use for this purpose receives little attention, and many analysts use the first indicator without considering whether there are better choices. When all indicators of the latent variable have essentially the same properties, then the choice matters less. But when this is not true, we could benefit from scaling indicator guidelines. Our article first demonstrates why latent variables need a scale. We then propose a set of criteria and accompanying diagnostic tools that can assist researchers in making informed decisions about scaling indicators. The criteria for a good scaling indicator include high face validity, high correlation with the latent variable, factor complexity of one, no correlated errors, no direct effects with other indicators, a minimal number of significant overidentification equation tests and modification indices, and invariance across groups and time. We demonstrate these criteria and diagnostics using two empirical examples and provide guidance on navigating conflicting results among criteria. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"868-889"},"PeriodicalIF":7.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10275390/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9650749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-03-27DOI: 10.1037/met0000568
Se-Kang Kim
A data matrix, where rows represent persons and columns represent measured subtests, can be viewed as a stack of person profiles, as rows are actually person profiles of observed responses on column subtests. Profile analysis seeks to identify a small number of latent profiles from a large number of person response profiles to identify central response patterns, which are useful for assessing the strengths and weaknesses of individuals across multiple dimensions in domains of interest. Moreover, the latent profiles are mathematically proven to be summative profiles that linearly combine all person response profiles. Since person response profiles are confounded with profile level and response pattern, the level effect must be controlled when they are factorized to identify a latent (or summative) profile that carries the response pattern effect. However, when the level effect is dominant but uncontrolled, only a summative profile carrying the level effect would be considered statistically meaningful according to a traditional metric (e.g., eigenvalue ≥ 1) or parallel analysis results. Nevertheless, the response pattern effect among individuals can provide assessment-relevant insights that are overlooked by conventional analysis; to achieve this, the level effect must be controlled. Consequently, the purpose of this study is to demonstrate how to correctly identify summative profiles containing central response patterns regardless of the centering techniques used on data sets. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"Factorization of person response profiles to identify summative profiles carrying central response patterns.","authors":"Se-Kang Kim","doi":"10.1037/met0000568","DOIUrl":"10.1037/met0000568","url":null,"abstract":"<p><p>A data matrix, where rows represent persons and columns represent measured subtests, can be viewed as a stack of person profiles, as rows are actually person profiles of observed responses on column subtests. Profile analysis seeks to identify a small number of latent profiles from a large number of person response profiles to identify central response patterns, which are useful for assessing the strengths and weaknesses of individuals across multiple dimensions in domains of interest. Moreover, the latent profiles are mathematically proven to be summative profiles that linearly combine all person response profiles. Since person response profiles are confounded with profile level and response pattern, the level effect must be controlled when they are factorized to identify a latent (or summative) profile that carries the response pattern effect. However, when the level effect is dominant but uncontrolled, only a summative profile carrying the level effect would be considered statistically meaningful according to a traditional metric (e.g., eigenvalue ≥ 1) or parallel analysis results. Nevertheless, the response pattern effect among individuals can provide assessment-relevant insights that are overlooked by conventional analysis; to achieve this, the level effect must be controlled. Consequently, the purpose of this study is to demonstrate how to correctly identify summative profiles containing central response patterns regardless of the centering techniques used on data sets. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"723-730"},"PeriodicalIF":7.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10016289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-06-12DOI: 10.1037/met0000585
Anja F Ernst, Casper J Albers, Marieke E Timmerman
Across different fields of research, the similarities and differences between various longitudinal models are not always eminently clear due to differences in data structure, application area, and terminology. Here we propose a comprehensive model framework that will allow simple comparisons between longitudinal models, to ease their empirical application and interpretation. At the within-individual level, our model framework accounts for various attributes of longitudinal data, such as growth and decline, cyclical trends, and the dynamic interplay between variables over time. At the between-individual level, our framework contains continuous and categorical latent variables to account for between-individual differences. This framework encompasses several well-known longitudinal models, including multilevel regression models, growth curve models, growth mixture models, vector-autoregressive models, and multilevel vector-autoregressive models. The general model framework is specified and its key characteristics are illustrated using famous longitudinal models as concrete examples. Various longitudinal models are reviewed and it is shown that all these models can be united into our comprehensive model framework. Extensions to the model framework are discussed. Recommendations for selecting and specifying longitudinal models are made for empirical researchers who aim to account for between-individual differences. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
在不同的研究领域,由于数据结构、应用领域和术语的不同,各种纵向模型之间的异同并不总是那么明显。在此,我们提出一个全面的模型框架,以便对纵向模型进行简单的比较,从而简化模型的实证应用和解释。在个体内部层面,我们的模型框架考虑了纵向数据的各种属性,如增长和下降、周期性趋势以及变量之间随时间变化的动态相互作用。在个体间层面,我们的框架包含连续和分类潜变量,以解释个体间的差异。这一框架包含多个著名的纵向模型,包括多层次回归模型、增长曲线模型、增长混合模型、向量自回归模型和多层次向量自回归模型。以著名的纵向模型为具体实例,说明了一般模型框架及其主要特征。综述了各种纵向模型,并表明所有这些模型都可以统一到我们的综合模型框架中。讨论了模型框架的扩展。为旨在考虑个体间差异的实证研究人员提供了选择和指定纵向模型的建议。(PsycInfo Database Record (c) 2024 APA, all rights reserved)。
{"title":"A comprehensive model framework for between-individual differences in longitudinal data.","authors":"Anja F Ernst, Casper J Albers, Marieke E Timmerman","doi":"10.1037/met0000585","DOIUrl":"10.1037/met0000585","url":null,"abstract":"<p><p>Across different fields of research, the similarities and differences between various longitudinal models are not always eminently clear due to differences in data structure, application area, and terminology. Here we propose a comprehensive model framework that will allow simple comparisons between longitudinal models, to ease their empirical application and interpretation. At the within-individual level, our model framework accounts for various attributes of longitudinal data, such as growth and decline, cyclical trends, and the dynamic interplay between variables over time. At the between-individual level, our framework contains continuous and categorical latent variables to account for between-individual differences. This framework encompasses several well-known longitudinal models, including multilevel regression models, growth curve models, growth mixture models, vector-autoregressive models, and multilevel vector-autoregressive models. The general model framework is specified and its key characteristics are illustrated using famous longitudinal models as concrete examples. Various longitudinal models are reviewed and it is shown that all these models can be united into our comprehensive model framework. Extensions to the model framework are discussed. Recommendations for selecting and specifying longitudinal models are made for empirical researchers who aim to account for between-individual differences. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"748-766"},"PeriodicalIF":7.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9612872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-05-11DOI: 10.1037/met0000582
Udo Boehm, Nathan J Evans, Quentin F Gronau, Dora Matzke, Eric-Jan Wagenmakers, Andrew J Heathcote
Cognitive models provide a substantively meaningful quantitative description of latent cognitive processes. The quantitative formulation of these models supports cumulative theory building and enables strong empirical tests. However, the nonlinearity of these models and pervasive correlations among model parameters pose special challenges when applying cognitive models to data. Firstly, estimating cognitive models typically requires large hierarchical data sets that need to be accommodated by an appropriate statistical structure within the model. Secondly, statistical inference needs to appropriately account for model uncertainty to avoid overconfidence and biased parameter estimates. In the present work, we show how these challenges can be addressed through a combination of Bayesian hierarchical modeling and Bayesian model averaging. To illustrate these techniques, we apply the popular diffusion decision model to data from a collaborative selective influence study. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
认知模型为潜在的认知过程提供了有实质意义的定量描述。这些模型的定量表述有助于累积理论的建立,并能进行强有力的实证检验。然而,这些模型的非线性以及模型参数之间普遍存在的相关性,给认知模型的数据应用带来了特殊的挑战。首先,认知模型的估算通常需要大量的分层数据集,而这些数据集需要通过模型内适当的统计结构来适应。其次,统计推断需要适当考虑模型的不确定性,以避免过度自信和有偏差的参数估计。在本研究中,我们展示了如何通过结合贝叶斯分层建模和贝叶斯模型平均来应对这些挑战。为了说明这些技术,我们将流行的扩散决策模型应用于一项合作选择性影响研究的数据中。(PsycInfo Database Record (c) 2024 APA, 版权所有)。
{"title":"Inclusion Bayes factors for mixed hierarchical diffusion decision models.","authors":"Udo Boehm, Nathan J Evans, Quentin F Gronau, Dora Matzke, Eric-Jan Wagenmakers, Andrew J Heathcote","doi":"10.1037/met0000582","DOIUrl":"10.1037/met0000582","url":null,"abstract":"<p><p>Cognitive models provide a substantively meaningful quantitative description of latent cognitive processes. The quantitative formulation of these models supports cumulative theory building and enables strong empirical tests. However, the nonlinearity of these models and pervasive correlations among model parameters pose special challenges when applying cognitive models to data. Firstly, estimating cognitive models typically requires large hierarchical data sets that need to be accommodated by an appropriate statistical structure within the model. Secondly, statistical inference needs to appropriately account for model uncertainty to avoid overconfidence and biased parameter estimates. In the present work, we show how these challenges can be addressed through a combination of Bayesian hierarchical modeling and Bayesian model averaging. To illustrate these techniques, we apply the popular diffusion decision model to data from a collaborative selective influence study. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"625-655"},"PeriodicalIF":7.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9796969","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-02-16DOI: 10.1037/met0000558
Guillermo Vallejo, María Paula Fernández, Pablo Esteban Livacic-Rojas
This article discusses the robustness of the multivariate analysis of covariance (MANCOVA) test for an emergent variable system and proposes a modification of this test to obtain adequate information from heterogeneous normal observations. The proposed approach for testing potential effects in heterogeneous MANCOVA models can be adopted effectively, regardless of the degree of heterogeneity and sample size imbalance. As our method was not designed to handle missing values, we also show how to derive the formulas for pooling the results of multiple-imputation-based analyses into a single final estimate. Results of simulated studies and analysis of real-data show that the proposed combining rules provide adequate coverage and power. Based on the current evidence, the two solutions suggested could be effectively used by researchers for testing hypotheses, provided that the data conform to normality. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
本文讨论了新兴变量系统的多变量协方差分析(MANCOVA)检验的稳健性,并提出了对该检验的一种修改方法,以便从异质正态观测中获取足够的信息。无论异质性和样本量不平衡的程度如何,都可以有效地采用所提出的方法来检验异质性 MANCOVA 模型中的潜在效应。由于我们的方法并不是为了处理缺失值而设计的,因此我们还展示了如何推导出将基于多重输入的分析结果汇集成一个最终估计值的公式。模拟研究和真实数据分析的结果表明,建议的合并规则具有足够的覆盖范围和能力。根据目前的证据,只要数据符合正态性,研究人员可以有效地使用所建议的两种解决方案来检验假设。(PsycInfo Database Record (c) 2023 APA, 版权所有)。
{"title":"Multivariate analysis of covariance for heterogeneous and incomplete data.","authors":"Guillermo Vallejo, María Paula Fernández, Pablo Esteban Livacic-Rojas","doi":"10.1037/met0000558","DOIUrl":"10.1037/met0000558","url":null,"abstract":"<p><p>This article discusses the robustness of the multivariate analysis of covariance (MANCOVA) test for an emergent variable system and proposes a modification of this test to obtain adequate information from heterogeneous normal observations. The proposed approach for testing potential effects in heterogeneous MANCOVA models can be adopted effectively, regardless of the degree of heterogeneity and sample size imbalance. As our method was not designed to handle missing values, we also show how to derive the formulas for pooling the results of multiple-imputation-based analyses into a single final estimate. Results of simulated studies and analysis of real-data show that the proposed combining rules provide adequate coverage and power. Based on the current evidence, the two solutions suggested could be effectively used by researchers for testing hypotheses, provided that the data conform to normality. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"731-747"},"PeriodicalIF":7.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10787830","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-04-13DOI: 10.1037/met0000569
Jillian C Strayhorn, Linda M Collins, David J Vanness
In current practice, intervention scientists applying the multiphase optimization strategy (MOST) with a 2k factorial optimization trial use a component screening approach (CSA) to select intervention components for inclusion in an optimized intervention. In this approach, scientists review all estimated main effects and interactions to identify the important ones based on a fixed threshold, and then base decisions about component selection on these important effects. We propose an alternative posterior expected value approach based on Bayesian decision theory. This new approach aims to be easier to apply and more readily extensible to a variety of intervention optimization problems. We used Monte Carlo simulation to evaluate the performance of a posterior expected value approach and CSA (automated for simulation purposes) relative to two benchmarks: random component selection, and the classical treatment package approach. We found that both the posterior expected value approach and CSA yielded substantial performance gains relative to the benchmarks. We also found that the posterior expected value approach outperformed CSA modestly but consistently in terms of overall accuracy, sensitivity, and specificity, across a wide range of realistic variations in simulated factorial optimization trials. We discuss implications for intervention optimization and promising future directions in the use of posterior expected value to make decisions in MOST. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
在目前的实践中,干预科学家在应用多阶段优化策略(MOST)进行 2k 因式优化试验时,会使用成分筛选法(CSA)来选择干预成分,以便将其纳入优化干预中。在这种方法中,科学家们会审查所有估计的主效应和交互作用,根据固定阈值确定重要效应,然后根据这些重要效应来决定干预成分的选择。我们提出了另一种基于贝叶斯决策理论的后验预期值方法。这种新方法更易于应用,也更容易扩展到各种干预优化问题中。我们使用蒙特卡罗模拟评估了后验期望值方法和 CSA(为模拟目的而自动进行)相对于随机成分选择和经典治疗包方法这两个基准的性能。我们发现,与基准相比,后验期望值法和 CSA 都能大幅提高性能。我们还发现,在模拟因子优化试验的各种现实变化中,后验期望值方法在总体准确性、灵敏度和特异性方面略微优于 CSA,但表现一致。我们讨论了干预优化的意义,以及使用后验预期值在 MOST 中做出决策的未来发展方向。(PsycInfo Database Record (c) 2023 APA, all rights reserved)。
{"title":"A posterior expected value approach to decision-making in the multiphase optimization strategy for intervention science.","authors":"Jillian C Strayhorn, Linda M Collins, David J Vanness","doi":"10.1037/met0000569","DOIUrl":"10.1037/met0000569","url":null,"abstract":"<p><p>In current practice, intervention scientists applying the multiphase optimization strategy (MOST) with a 2<i><sup>k</sup></i> factorial optimization trial use a component screening approach (CSA) to select intervention components for inclusion in an optimized intervention. In this approach, scientists review all estimated main effects and interactions to identify the important ones based on a fixed threshold, and then base decisions about component selection on these important effects. We propose an alternative posterior expected value approach based on Bayesian decision theory. This new approach aims to be easier to apply and more readily extensible to a variety of intervention optimization problems. We used Monte Carlo simulation to evaluate the performance of a posterior expected value approach and CSA (automated for simulation purposes) relative to two benchmarks: random component selection, and the classical treatment package approach. We found that both the posterior expected value approach and CSA yielded substantial performance gains relative to the benchmarks. We also found that the posterior expected value approach outperformed CSA modestly but consistently in terms of overall accuracy, sensitivity, and specificity, across a wide range of realistic variations in simulated factorial optimization trials. We discuss implications for intervention optimization and promising future directions in the use of posterior expected value to make decisions in MOST. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"656-678"},"PeriodicalIF":7.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9367545","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}