The development and evaluation of text classifiers in psychology depends on rigorous manual coding. Yet, the evaluation of manual coding and computational algorithms is usually considered separately. This is problematic because developing high-validity classifiers is a repeated process of identifying, explaining, and addressing conceptual and measurement issues during both the manual coding and classifier development stages. To address this problem, we introduce the Repeated Adjustment of Measurement Protocols (RAMP) method for developing high-validity text classifiers in psychology. The RAMP method has three stages: manual coding, classifier development, and integrative evaluation. These stages integrate the best practices of content analysis (manual coding), data science (classifier development), and psychology (integrative evaluation). Central to this integration is the concept of an inference loop, defined as the process of maximizing validity through repeated adjustments to concepts and constructs, guided by push-back from the empirical data. Inference loops operate both within each stage of the method and across related studies. We illustrate RAMP through a case study, where we manually coded 21,815 sentences for misunderstanding (Krippendorff's α = .79), and developed a rule-based classifier (Matthews correlation coefficient [MCC] = 0.22), a supervised machine learning classifier (Bidirectional Encoder Representations From Transformers; MCC = 0.69) and a large language model classifier (GPT-4o; MCC = 0.47). By integrating manual coding and classifier development stages, we were able to identify and address a concept validity problem with misunderstandings. RAMP advances existing methods by operationalizing validity as an ongoing dynamic process, where concepts and constructs are repeatedly adjusted toward increasingly widespread intersubjective agreement on their utility. (PsycInfo Database Record (c) 2026 APA, all rights reserved).
{"title":"The repeated adjustment of measurement protocols method for developing high-validity text classifiers.","authors":"Alex Goddard, Alex Gillespie","doi":"10.1037/met0000787","DOIUrl":"10.1037/met0000787","url":null,"abstract":"<p><p>The development and evaluation of text classifiers in psychology depends on rigorous manual coding. Yet, the evaluation of manual coding and computational algorithms is usually considered separately. This is problematic because developing high-validity classifiers is a repeated process of identifying, explaining, and addressing conceptual and measurement issues during both the manual coding and classifier development stages. To address this problem, we introduce the Repeated Adjustment of Measurement Protocols (RAMP) method for developing high-validity text classifiers in psychology. The RAMP method has three stages: manual coding, classifier development, and integrative evaluation. These stages integrate the best practices of content analysis (manual coding), data science (classifier development), and psychology (integrative evaluation). Central to this integration is the concept of an inference loop, defined as the process of maximizing validity through repeated adjustments to concepts and constructs, guided by push-back from the empirical data. Inference loops operate both within each stage of the method and across related studies. We illustrate RAMP through a case study, where we manually coded 21,815 sentences for misunderstanding (Krippendorff's α = .79), and developed a rule-based classifier (Matthews correlation coefficient [MCC] = 0.22), a supervised machine learning classifier (Bidirectional Encoder Representations From Transformers; MCC = 0.69) and a large language model classifier (GPT-4o; MCC = 0.47). By integrating manual coding and classifier development stages, we were able to identify and address a concept validity problem with misunderstandings. RAMP advances existing methods by operationalizing validity as an ongoing dynamic process, where concepts and constructs are repeatedly adjusted toward increasingly widespread intersubjective agreement on their utility. (PsycInfo Database Record (c) 2026 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":""},"PeriodicalIF":7.8,"publicationDate":"2025-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145233362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Supplemental Material for Causal Mediation Analysis With Two Mediators: A Comprehensive Guide to Estimating Total and Natural Effects Across Various Multiple Mediators Setups","authors":"","doi":"10.1037/met0000781.supp","DOIUrl":"https://doi.org/10.1037/met0000781.supp","url":null,"abstract":"","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":"55 1","pages":""},"PeriodicalIF":7.0,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145254622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ivan Jacob Agaloos Pesigan, Michael A Russell, Sy-Miin Chow
Mediation modeling using longitudinal data is an exciting field that captures the interrelations in dynamic changes, such as mediated changes, over time. Even though discrete-time vector autoregressive approaches are commonly used to estimate indirect effects in longitudinal data, they have known limitations due to the dependency of inferential results on the time intervals between successive occasions and the assumption of regular spacing between measurements. Continuous-time vector autoregressive models have been proposed as an alternative to address these issues. Previous work in the area (e.g., Deboeck & Preacher, 2015; Ryan & Hamaker, 2021) has shown how the direct, indirect, and total effects, for a range of time-interval values, can be calculated using parameters estimated from continuous-time vector autoregressive models for causal inferential purposes. However, both standardized effects size measures and methods for calculating the uncertainty around the direct, indirect, and total effects in continuous-time mediation have yet to be explored. Drawing from the mediation model literature, we present and compare results using the delta, Monte Carlo, and parametric bootstrap methods to calculate SEs and confidence intervals for the direct, indirect, and total effects in continuous-time mediation for inferential purposes. Options to automate these inferential procedures and facilitate interpretations are available in the cTMed R package. (PsycInfo Database Record (c) 2026 APA, all rights reserved).
使用纵向数据的中介建模是一个令人兴奋的领域,它可以捕获动态变化(如中介变化)中随时间变化的相互关系。尽管离散时间向量自回归方法通常用于估计纵向数据中的间接影响,但由于推论结果依赖于连续事件之间的时间间隔和假设测量之间的规则间隔,它们具有已知的局限性。连续时间向量自回归模型已被提出作为解决这些问题的替代方案。该领域之前的研究(例如,Deboeck & Preacher, 2015; Ryan & Hamaker, 2021)表明,为了进行因果推理,可以使用从连续时间向量自回归模型中估计的参数来计算一系列时间间隔值的直接、间接和总效应。然而,对于连续时间中介中直接、间接和总效应的不确定性的计算方法和标准化效应大小的测量方法还有待探索。根据中介模型文献,我们提出并比较了使用delta、Monte Carlo和参数自举方法的结果,以计算连续时间中介中直接、间接和总效应的se和置信区间。cTMed R包中提供了自动化这些推理过程和促进解释的选项。(PsycInfo Database Record (c) 2025 APA,版权所有)。
{"title":"Inferences and effect sizes for direct, indirect, and total effects in continuous-time mediation models.","authors":"Ivan Jacob Agaloos Pesigan, Michael A Russell, Sy-Miin Chow","doi":"10.1037/met0000779","DOIUrl":"10.1037/met0000779","url":null,"abstract":"<p><p>Mediation modeling using longitudinal data is an exciting field that captures the interrelations in dynamic changes, such as mediated changes, over time. Even though discrete-time vector autoregressive approaches are commonly used to estimate indirect effects in longitudinal data, they have known limitations due to the dependency of inferential results on the time intervals between successive occasions and the assumption of regular spacing between measurements. Continuous-time vector autoregressive models have been proposed as an alternative to address these issues. Previous work in the area (e.g., Deboeck & Preacher, 2015; Ryan & Hamaker, 2021) has shown how the direct, indirect, and total effects, for a range of time-interval values, can be calculated using parameters estimated from continuous-time vector autoregressive models for causal inferential purposes. However, both standardized effects size measures and methods for calculating the uncertainty around the direct, indirect, and total effects in continuous-time mediation have yet to be explored. Drawing from the mediation model literature, we present and compare results using the delta, Monte Carlo, and parametric bootstrap methods to calculate SEs and confidence intervals for the direct, indirect, and total effects in continuous-time mediation for inferential purposes. Options to automate these inferential procedures and facilitate interpretations are available in the cTMed R package. (PsycInfo Database Record (c) 2026 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":""},"PeriodicalIF":7.8,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12494154/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145213089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Supplemental Material for The Repeated Adjustment of Measurement Protocols Method for Developing High-Validity Text Classifiers","authors":"","doi":"10.1037/met0000787.supp","DOIUrl":"https://doi.org/10.1037/met0000787.supp","url":null,"abstract":"","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":"29 1","pages":""},"PeriodicalIF":7.0,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145254633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2023-07-20DOI: 10.1037/met0000599
David Trafimow, Michael R Hyman, Alena Kostyk
The generally small but touted as "statistically significant" correlation coefficients in the social sciences jeopardize theory testing and prediction. To investigate these small coefficients' underlying causes, traditional equations such as Spearman's (1904) classic attenuation formula, Cronbach's (1951) alpha, and Guilford and Fruchter's (1973) equation for the effect of additional items on a scale's predictive power are considered. These equations' implications differ regarding large interitem correlations enhancing or diminishing predictive power. Contrary to conventional practice, such correlations decrease predictive power when treating items as multi-item scale components but can increase predictive power when treating items separately. The implications are wide-ranging. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
在社会科学中,通常较小但被吹捧为“统计显著”的相关系数危及理论检验和预测。为了研究这些小系数的潜在原因,考虑了传统方程,如Spearman(1904)的经典衰减公式,Cronbach(1951)的alpha,以及Guilford和Fruchter(1973)的附加项对量表预测能力影响的方程。这些方程的含义不同于大的项目间相关性,增强或减弱预测能力。与传统做法相反,当将项目作为多项目量表组件处理时,这种相关性会降低预测能力,但当单独处理项目时,这种相关性会增加预测能力。其影响是广泛的。(PsycInfo Database Record (c) 2025 APA,版权所有)。
{"title":"Enhancing predictive power by unamalgamating multi-item scales.","authors":"David Trafimow, Michael R Hyman, Alena Kostyk","doi":"10.1037/met0000599","DOIUrl":"10.1037/met0000599","url":null,"abstract":"<p><p>The generally small but touted as \"statistically significant\" correlation coefficients in the social sciences jeopardize theory testing and prediction. To investigate these small coefficients' underlying causes, traditional equations such as Spearman's (1904) classic attenuation formula, Cronbach's (1951) alpha, and Guilford and Fruchter's (1973) equation for the effect of additional items on a scale's predictive power are considered. These equations' implications differ regarding large interitem correlations enhancing or diminishing predictive power. Contrary to conventional practice, such correlations decrease predictive power when treating items as multi-item scale components but can increase predictive power when treating items separately. The implications are wide-ranging. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"1043-1055"},"PeriodicalIF":7.8,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9838250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2023-09-07DOI: 10.1037/met0000602
Irene Klugkist, Thom Benjamin Volker
To establish a theory one needs cleverly designed and well-executed studies with appropriate and correctly interpreted statistical analyses. Equally important, one also needs replications of such studies and a way to combine the results of several replications into an accumulated state of knowledge. An approach that provides an appropriate and powerful analysis for studies targeting prespecified theories is the use of Bayesian informative hypothesis testing. An additional advantage of the use of this Bayesian approach is that combining the results from multiple studies is straightforward. In this article, we discuss the behavior of Bayes factors in the context of evaluating informative hypotheses with multiple studies. By using simple models and (partly) analytical solutions, we introduce and evaluate Bayesian evidence synthesis (BES) and compare its results to Bayesian sequential updating. By doing so, we clarify how different replications or updating questions can be evaluated. In addition, we illustrate BES with two simulations, in which multiple studies are generated to resemble conceptual replications. The studies in these simulations are too heterogeneous to be aggregated with conventional research synthesis methods. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
为了建立一个理论,人们需要巧妙设计和良好执行的研究,以及适当和正确解释的统计分析。同样重要的是,人们还需要重复这样的研究,并找到一种方法,将多次重复的结果结合起来,形成一种积累的知识状态。使用贝叶斯信息假设检验是一种为针对预先指定理论的研究提供适当和有力分析的方法。使用贝叶斯方法的另一个优点是,将多个研究的结果结合起来是直接的。在这篇文章中,我们讨论了贝叶斯因素的行为在评估信息假设与多个研究的背景下。通过使用简单的模型和(部分)解析解,我们介绍和评估了贝叶斯证据合成(BES),并将其结果与贝叶斯序列更新进行了比较。通过这样做,我们阐明了如何评估不同的重复或更新问题。此外,我们用两个模拟来说明BES,其中生成了多个类似概念复制的研究。这些模拟中的研究太过异质,无法用传统的研究综合方法进行汇总。(PsycInfo Database Record (c) 2025 APA,版权所有)。
{"title":"Bayesian evidence synthesis for informative hypotheses: An introduction.","authors":"Irene Klugkist, Thom Benjamin Volker","doi":"10.1037/met0000602","DOIUrl":"10.1037/met0000602","url":null,"abstract":"<p><p>To establish a theory one needs cleverly designed and well-executed studies with appropriate and correctly interpreted statistical analyses. Equally important, one also needs replications of such studies and a way to combine the results of several replications into an accumulated state of knowledge. An approach that provides an appropriate and powerful analysis for studies targeting prespecified theories is the use of Bayesian informative hypothesis testing. An additional advantage of the use of this Bayesian approach is that combining the results from multiple studies is straightforward. In this article, we discuss the behavior of Bayes factors in the context of evaluating informative hypotheses with multiple studies. By using simple models and (partly) analytical solutions, we introduce and evaluate Bayesian evidence synthesis (BES) and compare its results to Bayesian sequential updating. By doing so, we clarify how different replications or updating questions can be evaluated. In addition, we illustrate BES with two simulations, in which multiple studies are generated to resemble conceptual replications. The studies in these simulations are too heterogeneous to be aggregated with conventional research synthesis methods. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"949-965"},"PeriodicalIF":7.8,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10173540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2023-12-25DOI: 10.1037/met0000624
Esther Maassen, E Damiano D'Urso, Marcel A L M van Assen, Michèle B Nuijten, Kim De Roover, Jelte M Wicherts
Self-report scales are widely used in psychology to compare means in latent constructs across groups, experimental conditions, or time points. However, for these comparisons to be meaningful and unbiased, the scales must demonstrate measurement invariance (MI) across compared time points or (experimental) groups. MI testing determines whether the latent constructs are measured equivalently across groups or time, which is essential for meaningful comparisons. We conducted a systematic review of 426 psychology articles with openly available data, to (a) examine common practices in conducting and reporting of MI testing, (b) assess whether we could reproduce the reported MI results, and (c) conduct MI tests for the comparisons that enabled sufficiently powerful MI testing. We identified 96 articles that contained a total of 929 comparisons. Results showed that only 4% of the 929 comparisons underwent MI testing, and the tests were generally poorly reported. None of the reported MI tests were reproducible, and only 26% of the 174 newly performed MI tests reached sufficient (scalar) invariance, with MI failing completely in 58% of tests. Exploratory analyses suggested that in nearly half of the comparisons where configural invariance was rejected, the number of factors differed between groups. These results indicate that MI tests are rarely conducted and poorly reported in psychological studies. We observed frequent violations of MI, suggesting that reported differences between (experimental) groups may not be solely attributed to group differences in the latent constructs. We offer recommendations aimed at improving reporting and computational reproducibility practices in psychology. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
在心理学中,自我报告量表被广泛用于比较不同组别、实验条件或时间点的潜在结构的平均值。然而,要使这些比较有意义且无偏见,量表必须在比较的时间点或(实验)组间表现出测量不变性(MI)。测量不变性测试可确定各组或各时间点对潜构的测量是否等效,这对于进行有意义的比较至关重要。我们对 426 篇公开数据的心理学文章进行了系统性回顾,目的是:(a)检查进行和报告 MI 检验的常见做法;(b)评估我们是否能重现所报告的 MI 结果;以及(c)对能进行足够强大 MI 检验的比较进行 MI 检验。我们确定了 96 篇文章,共包含 929 项比较。结果显示,在 929 项比较中,只有 4% 的比较进行了多元智能测试,而且这些测试的报道普遍较少。所报道的 MI 测试都不具有可重复性,在 174 项新进行的 MI 测试中,只有 26% 达到了足够的(标度)不变性,58% 的测试完全不符合 MI 标准。探索性分析表明,在配置不变量被拒绝的近一半比较中,各组之间的因子数量存在差异。这些结果表明,在心理学研究中,多元智能测试很少进行,报告也很少。我们观察到经常出现违反多元智能的情况,这表明所报告的(实验)组间差异可能并不完全归因于潜在建构的组间差异。我们提出了旨在改进心理学报告和计算可重复性实践的建议。(PsycInfo Database Record (c) 2023 APA, 版权所有)。
{"title":"The dire disregard of measurement invariance testing in psychological science.","authors":"Esther Maassen, E Damiano D'Urso, Marcel A L M van Assen, Michèle B Nuijten, Kim De Roover, Jelte M Wicherts","doi":"10.1037/met0000624","DOIUrl":"10.1037/met0000624","url":null,"abstract":"<p><p>Self-report scales are widely used in psychology to compare means in latent constructs across groups, experimental conditions, or time points. However, for these comparisons to be meaningful and unbiased, the scales must demonstrate measurement invariance (MI) across compared time points or (experimental) groups. MI testing determines whether the latent constructs are measured equivalently across groups or time, which is essential for meaningful comparisons. We conducted a systematic review of 426 psychology articles with openly available data, to (a) examine common practices in conducting and reporting of MI testing, (b) assess whether we could reproduce the reported MI results, and (c) conduct MI tests for the comparisons that enabled sufficiently powerful MI testing. We identified 96 articles that contained a total of 929 comparisons. Results showed that only 4% of the 929 comparisons underwent MI testing, and the tests were generally poorly reported. None of the reported MI tests were reproducible, and only 26% of the 174 newly performed MI tests reached sufficient (scalar) invariance, with MI failing completely in 58% of tests. Exploratory analyses suggested that in nearly half of the comparisons where configural invariance was rejected, the number of factors differed between groups. These results indicate that MI tests are rarely conducted and poorly reported in psychological studies. We observed frequent violations of MI, suggesting that reported differences between (experimental) groups may not be solely attributed to group differences in the latent constructs. We offer recommendations aimed at improving reporting and computational reproducibility practices in psychology. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"966-979"},"PeriodicalIF":7.8,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139037948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2023-10-16DOI: 10.1037/met0000615
Mar J F Ollero, Eduardo Estrada, Michael D Hunter, Pablo F Cáncer
People show stable differences in the way their affect fluctuates over time. Within the general framework of dynamical systems, the damped linear oscillator (DLO) model has been proposed as a useful approach to study affect dynamics. The DLO model can be applied to repeated measures provided by a single individual, and the resulting parameters can capture relevant features of the person's affect dynamics. Focusing on negative affect, we provide an accessible interpretation of the DLO model parameters in terms of emotional lability, resilience, and vulnerability. We conducted a Monte Carlo study to test the DLO model performance under different empirically relevant conditions in terms of individual characteristics and sampling scheme. We used state-space models in continuous time. The results show that, under certain conditions, the DLO model is able to accurately and efficiently recover the parameters underlying the affective dynamics of a single individual. We discuss the results and the theoretical and practical implications of using this model, illustrate how to use it for studying psychological phenomena at the individual level, and provide specific recommendations on how to collect data for this purpose. We also provide a tutorial website and computer code in R to implement this approach. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
{"title":"Characterizing affect dynamics with a damped linear oscillator model: Theoretical considerations and recommendations for individual-level applications.","authors":"Mar J F Ollero, Eduardo Estrada, Michael D Hunter, Pablo F Cáncer","doi":"10.1037/met0000615","DOIUrl":"10.1037/met0000615","url":null,"abstract":"<p><p>People show stable differences in the way their affect fluctuates over time. Within the general framework of dynamical systems, the damped linear oscillator (DLO) model has been proposed as a useful approach to study affect dynamics. The DLO model can be applied to repeated measures provided by a single individual, and the resulting parameters can capture relevant features of the person's affect dynamics. Focusing on negative affect, we provide an accessible interpretation of the DLO model parameters in terms of emotional lability, resilience, and vulnerability. We conducted a Monte Carlo study to test the DLO model performance under different empirically relevant conditions in terms of individual characteristics and sampling scheme. We used state-space models in continuous time. The results show that, under certain conditions, the DLO model is able to accurately and efficiently recover the parameters underlying the affective dynamics of a single individual. We discuss the results and the theoretical and practical implications of using this model, illustrate how to use it for studying psychological phenomena at the individual level, and provide specific recommendations on how to collect data for this purpose. We also provide a tutorial website and computer code in R to implement this approach. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"1095-1112"},"PeriodicalIF":7.8,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41238100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2023-08-10DOI: 10.1037/met0000605
Miriam Brinberg, Graham D Bodie, Denise H Solomon, Susanne M Jones, Nilam Ram
Several theoretical perspectives suggest that dyadic experiences are distinguished by patterns of behavioral change that emerge during interactions. Methods for examining change in behavior over time are well elaborated for the study of change along continuous dimensions. Extensions for charting increases and decreases in individuals' use of specific, categorically defined behaviors, however, are rarely invoked. Greater accessibility of Bayesian frameworks that facilitate formulation and estimation of the requisite models is opening new possibilities. This article provides a primer on how multinomial logistic growth models can be used to examine between-dyad differences in within-dyad behavioral change over the course of an interaction. We describe and illustrate how these models are implemented in the Bayesian framework using data from support conversations between strangers (N = 118 dyads) to examine (RQ1) how six types of listeners' and disclosers' behaviors change as support conversations unfold and (RQ2) how the disclosers' preconversation distress moderates the change in conversation behaviors. The primer concludes with a series of notes on (a) implications of modeling choices, (b) flexibility in modeling nonlinear change, (c) necessity for theory that specifies how and why change trajectories differ, and (d) how multinomial logistic growth models can help refine current theory about dyadic interaction. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
一些理论观点认为,二元体验是通过互动过程中出现的行为变化模式来区分的。研究行为随时间变化的方法在研究连续维度的变化方面得到了很好的阐述。然而,在个人使用特定的、分类定义的行为时,图表扩展的增减很少被调用。贝叶斯框架的更大可及性,促进了必要模型的制定和估计,正在开辟新的可能性。本文提供了如何使用多项逻辑增长模型来检查在相互作用过程中对内行为变化的对间差异的入门。我们描述并说明了这些模型是如何在贝叶斯框架中实现的,使用陌生人之间的支持对话(N = 118对)的数据来检查(RQ1)六种类型的听者和披露者的行为如何随着支持对话的展开而变化,(RQ2)披露者的谈话前痛苦如何调节谈话行为的变化。本导论以一系列关于(a)建模选择的含义,(b)建模非线性变化的灵活性,(c)指定变化轨迹如何以及为什么不同的理论的必要性,以及(d)多项逻辑增长模型如何帮助完善关于二元相互作用的当前理论。(PsycInfo Database Record (c) 2025 APA,版权所有)。
{"title":"Examining individual differences in how interaction behaviors change over time: A dyadic multinomial logistic growth modeling approach.","authors":"Miriam Brinberg, Graham D Bodie, Denise H Solomon, Susanne M Jones, Nilam Ram","doi":"10.1037/met0000605","DOIUrl":"10.1037/met0000605","url":null,"abstract":"<p><p>Several theoretical perspectives suggest that dyadic experiences are distinguished by patterns of behavioral change that emerge during interactions. Methods for examining change in behavior over time are well elaborated for the study of change along continuous dimensions. Extensions for charting increases and decreases in individuals' use of specific, categorically defined behaviors, however, are rarely invoked. Greater accessibility of Bayesian frameworks that facilitate formulation and estimation of the requisite models is opening new possibilities. This article provides a primer on how multinomial logistic growth models can be used to examine between-dyad differences in within-dyad behavioral change over the course of an interaction. We describe and illustrate how these models are implemented in the Bayesian framework using data from support conversations between strangers (<i>N</i> = 118 dyads) to examine (RQ1) how six types of listeners' and disclosers' behaviors change as support conversations unfold and (RQ2) how the disclosers' preconversation distress moderates the change in conversation behaviors. The primer concludes with a series of notes on (a) implications of modeling choices, (b) flexibility in modeling nonlinear change, (c) necessity for theory that specifies how and why change trajectories differ, and (d) how multinomial logistic growth models can help refine current theory about dyadic interaction. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"1079-1094"},"PeriodicalIF":7.8,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9967422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2023-12-21DOI: 10.1037/met0000630
Eunsook Kim, Yan Wang, Hsien-Yuan Hsu
Factor mixture modeling (FMM) incorporates both continuous latent variables and categorical latent variables in a single analytic model clustering items and observations simultaneously. After two decades since the introduction of FMM to psychological and behavioral science research, it is an opportune time to review FMM applications to understand how these applications are utilized in real-world research. We conducted a systematic review of 76 FMM applications. We developed a comprehensive coding scheme based on the current methodological literature of FMM and evaluated common usages and practices of FMM. Based on the review, we identify challenges and issues that applied researchers encounter in the practice of FMM and provide practical suggestions to promote well-informed decision making. Lastly, we discuss future methodological directions and suggest how FMM can be expanded beyond its typical use in applied studies. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
因子混合建模(FMM)将连续潜变量和分类潜变量纳入一个分析模型,同时对项目和观察结果进行聚类。自 FMM 被引入心理学和行为科学研究二十年以来,现在正是回顾 FMM 应用以了解这些应用在实际研究中的应用情况的大好时机。我们对 76 项 FMM 应用进行了系统回顾。我们根据当前的 FMM 方法论文献制定了一套全面的编码方案,并评估了 FMM 的常见用法和实践。在综述的基础上,我们确定了应用研究人员在 FMM 实践中遇到的挑战和问题,并提供了实用建议,以促进知情决策。最后,我们讨论了未来的方法论方向,并建议如何将 FMM 扩展到应用研究的典型用途之外。(PsycInfo Database Record (c) 2023 APA, all rights reserved)。
{"title":"A systematic review of and reflection on the applications of factor mixture modeling.","authors":"Eunsook Kim, Yan Wang, Hsien-Yuan Hsu","doi":"10.1037/met0000630","DOIUrl":"10.1037/met0000630","url":null,"abstract":"<p><p>Factor mixture modeling (FMM) incorporates both continuous latent variables and categorical latent variables in a single analytic model clustering items and observations simultaneously. After two decades since the introduction of FMM to psychological and behavioral science research, it is an opportune time to review FMM applications to understand how these applications are utilized in real-world research. We conducted a systematic review of 76 FMM applications. We developed a comprehensive coding scheme based on the current methodological literature of FMM and evaluated common usages and practices of FMM. Based on the review, we identify challenges and issues that applied researchers encounter in the practice of FMM and provide practical suggestions to promote well-informed decision making. Lastly, we discuss future methodological directions and suggest how FMM can be expanded beyond its typical use in applied studies. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":20782,"journal":{"name":"Psychological methods","volume":" ","pages":"997-1016"},"PeriodicalIF":7.8,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138831225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}