The Q-matrix is a crucial component of cognitive diagnostic theory and an important basis for the research and practical application of cognitive diagnosis. In practice, the Q-matrix is typically developed by domain experts and may contain some misspecifications, so it needs to be refined using Q-matrix validation methods. Based on signal detection theory, this paper puts forward a new Q-matrix validation method (i.e., method) and then conducts a simulation study to compare the new method with existing methods. The results show that when the model is DINA (deterministic inputs, noisy 'and' gate), the method outperforms the existing methods under all conditions; under the generalized DINA (G-DINA) model, the method still has the highest validation rate when the sample size is small, and the item quality is high or the rate of Q-matrix misspecification is ≥.4. Finally, a sub-dataset of the PISA 2000 reading assessment is analysed to evaluate the reliability of the method.
{"title":"A new Q-matrix validation method based on signal detection theory.","authors":"Jia Li, Ping Chen","doi":"10.1111/bmsp.12371","DOIUrl":"https://doi.org/10.1111/bmsp.12371","url":null,"abstract":"<p><p>The Q-matrix is a crucial component of cognitive diagnostic theory and an important basis for the research and practical application of cognitive diagnosis. In practice, the Q-matrix is typically developed by domain experts and may contain some misspecifications, so it needs to be refined using Q-matrix validation methods. Based on signal detection theory, this paper puts forward a new Q-matrix validation method (i.e., <math> <semantics><mrow><mi>β</mi></mrow> <annotation>$$ beta $$</annotation></semantics> </math> method) and then conducts a simulation study to compare the new method with existing methods. The results show that when the model is DINA (deterministic inputs, noisy 'and' gate), the <math> <semantics><mrow><mi>β</mi></mrow> <annotation>$$ beta $$</annotation></semantics> </math> method outperforms the existing methods under all conditions; under the generalized DINA (G-DINA) model, the method still has the highest validation rate when the sample size is small, and the item quality is high or the rate of Q-matrix misspecification is ≥.4. Finally, a sub-dataset of the PISA 2000 reading assessment is analysed to evaluate the reliability of the <math> <semantics><mrow><mi>β</mi></mrow> <annotation>$$ beta $$</annotation></semantics> </math> method.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142677821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The discriminability in polytomous KST was introduced by Stefanutti et al. (Journal of Mathematical Psychology, 2020, 94, 102306). As the interesting topic in polytomous KST, this paper discusses the discriminability around granular polytomous knowledge spaces, polytomous knowledge structures, polytomous surmising functions and polytomous skill functions. More precisely, this paper gives some equivalences between the discriminability of polytomous surmising functions (resp. polytomous skill functions) and the discriminability of granular polytomous knowledge spaces (resp. polytomous knowledge structures). Such findings open the field to a systematic generalization of the discriminability in KST to the polytomous case.
{"title":"Discriminability around polytomous knowledge structures and polytomous functions.","authors":"Xun Ge","doi":"10.1111/bmsp.12370","DOIUrl":"https://doi.org/10.1111/bmsp.12370","url":null,"abstract":"<p><p>The discriminability in polytomous KST was introduced by Stefanutti et al. (Journal of Mathematical Psychology, 2020, 94, 102306). As the interesting topic in polytomous KST, this paper discusses the discriminability around granular polytomous knowledge spaces, polytomous knowledge structures, polytomous surmising functions and polytomous skill functions. More precisely, this paper gives some equivalences between the discriminability of polytomous surmising functions (resp. polytomous skill functions) and the discriminability of granular polytomous knowledge spaces (resp. polytomous knowledge structures). Such findings open the field to a systematic generalization of the discriminability in KST to the polytomous case.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142649914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Interaction analysis using linear regression is widely employed in psychology and related fields, yet it often induces confusion among applied researchers and students. This paper aims to address this confusion by developing intuitive visual explanations based on causal graphs. By leveraging causal graphs with distinct interaction nodes, we provide clear insights into interpreting main effects in the presence of interaction, the rationale behind centering to reduce multicollinearity, and other pertinent topics. The proposed graphical approach could serve as a useful complement to existing algebraic explanations, fostering a more comprehensive understanding of the mechanics of linear interaction analysis.
{"title":"Understanding linear interaction analysis with causal graphs.","authors":"Yongnam Kim, Geryong Jung","doi":"10.1111/bmsp.12369","DOIUrl":"https://doi.org/10.1111/bmsp.12369","url":null,"abstract":"<p><p>Interaction analysis using linear regression is widely employed in psychology and related fields, yet it often induces confusion among applied researchers and students. This paper aims to address this confusion by developing intuitive visual explanations based on causal graphs. By leveraging causal graphs with distinct interaction nodes, we provide clear insights into interpreting main effects in the presence of interaction, the rationale behind centering to reduce multicollinearity, and other pertinent topics. The proposed graphical approach could serve as a useful complement to existing algebraic explanations, fostering a more comprehensive understanding of the mechanics of linear interaction analysis.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142633300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In addition to the usual slope and location parameters included in a regular two-parameter logistic model (2PL), the logistic positive exponent (LPE) model incorporates an item parameter that leads to asymmetric item characteristic curves, which have recently been shown to be useful in some contexts. Although this model has been used in some empirical studies, an identifiability analysis (i.e., checking the (un)identified status of a model and searching for identifiablity restrictions to make an unidentified model identified) has not yet been established. In this paper, we formalize the unidentified status of a large class of fixed-effects item response theory models that includes the LPE model and related versions of it. In addition, we conduct an identifiability analysis of a particular version of the LPE model that is based on the fixed-effects one-parameter logistic model (1PL), which we call the 1PL-LPE model. The main result indicates that the 1PL-LPE model is not identifiable. Ways to make the 1PL-LPE useful in practice and how different strategies for identifiability analyses may affect other versions of the model are also discussed.
{"title":"Identifiability analysis of the fixed-effects one-parameter logistic positive exponent model.","authors":"Jorge González, Jorge Bazán, Mariana Curi","doi":"10.1111/bmsp.12366","DOIUrl":"https://doi.org/10.1111/bmsp.12366","url":null,"abstract":"<p><p>In addition to the usual slope and location parameters included in a regular two-parameter logistic model (2PL), the logistic positive exponent (LPE) model incorporates an item parameter that leads to asymmetric item characteristic curves, which have recently been shown to be useful in some contexts. Although this model has been used in some empirical studies, an identifiability analysis (i.e., checking the (un)identified status of a model and searching for identifiablity restrictions to make an unidentified model identified) has not yet been established. In this paper, we formalize the unidentified status of a large class of fixed-effects item response theory models that includes the LPE model and related versions of it. In addition, we conduct an identifiability analysis of a particular version of the LPE model that is based on the fixed-effects one-parameter logistic model (1PL), which we call the 1PL-LPE model. The main result indicates that the 1PL-LPE model is not identifiable. Ways to make the 1PL-LPE useful in practice and how different strategies for identifiability analyses may affect other versions of the model are also discussed.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142633297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Q-matrices are crucial components of cognitive diagnosis models (CDMs), which are used to provide diagnostic information and classify examinees according to their attribute profiles. The absence of an appropriate Q-matrix that correctly reflects item-attribute relationships often limits the widespread use of CDMs. Rather than relying on expert judgment for specification and post-hoc methods for validation, there has been a notable shift towards Q-matrix estimation by adopting Bayesian methods. Nevertheless, their dependency on Markov chain Monte Carlo (MCMC) estimation requires substantial computational burdens and their exploratory tendency is unscalable to large-scale settings. As a scalable and efficient alternative, this study introduces the partially confirmatory framework within a saturated CDM, where the Q-matrix can be partially defined by experts and partially inferred from data. To address the dual needs of accuracy and efficiency, the proposed framework accommodates two estimation algorithms-an MCMC algorithm and a Variational Bayesian Expectation Maximization (VBEM) algorithm. This dual-channel approach extends the model's applicability across a variety of settings. Based on simulated and real data, the proposed framework demonstrated its robustness in Q-matrix inference.
{"title":"Regularized Bayesian algorithms for Q-matrix inference based on saturated cognitive diagnosis modelling.","authors":"Yi Jin, Jinsong Chen","doi":"10.1111/bmsp.12368","DOIUrl":"https://doi.org/10.1111/bmsp.12368","url":null,"abstract":"<p><p>Q-matrices are crucial components of cognitive diagnosis models (CDMs), which are used to provide diagnostic information and classify examinees according to their attribute profiles. The absence of an appropriate Q-matrix that correctly reflects item-attribute relationships often limits the widespread use of CDMs. Rather than relying on expert judgment for specification and post-hoc methods for validation, there has been a notable shift towards Q-matrix estimation by adopting Bayesian methods. Nevertheless, their dependency on Markov chain Monte Carlo (MCMC) estimation requires substantial computational burdens and their exploratory tendency is unscalable to large-scale settings. As a scalable and efficient alternative, this study introduces the partially confirmatory framework within a saturated CDM, where the Q-matrix can be partially defined by experts and partially inferred from data. To address the dual needs of accuracy and efficiency, the proposed framework accommodates two estimation algorithms-an MCMC algorithm and a Variational Bayesian Expectation Maximization (VBEM) algorithm. This dual-channel approach extends the model's applicability across a variety of settings. Based on simulated and real data, the proposed framework demonstrated its robustness in Q-matrix inference.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142633298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Item response tree (IRTree) models form a family of psychometric models that allow researchers to control for multiple response processes, such as different sorts of response styles, in the measurement of latent traits. While IRTree models can capture quantitative individual differences in both the latent traits of interest and the use of response categories, they maintain the basic assumption that the nature and weighting of latent response processes are homogeneous across the entire population of respondents. In the present research, we therefore propose a novel approach for detecting heterogeneity in the parameters of IRTree models across subgroups that engage in different response behavior. The approach uses score-based tests to reveal violations of parameter heterogeneity along extraneous person covariates, and it can be employed as a model-based partitioning algorithm to identify sources of differences in the strength of trait-based responding or other response processes. Simulation studies demonstrate generally accurate Type I error rates and sufficient power for metric, ordinal, and categorical person covariates and for different types of test statistics, with the potential to differentiate between different types of parameter heterogeneity. An empirical application illustrates the use of score-based partitioning in the analysis of latent response processes with real data.
{"title":"Investigating heterogeneity in IRTree models for multiple response processes with score-based partitioning.","authors":"Rudolf Debelak, Thorsten Meiser, Alicia Gernand","doi":"10.1111/bmsp.12367","DOIUrl":"https://doi.org/10.1111/bmsp.12367","url":null,"abstract":"<p><p>Item response tree (IRTree) models form a family of psychometric models that allow researchers to control for multiple response processes, such as different sorts of response styles, in the measurement of latent traits. While IRTree models can capture quantitative individual differences in both the latent traits of interest and the use of response categories, they maintain the basic assumption that the nature and weighting of latent response processes are homogeneous across the entire population of respondents. In the present research, we therefore propose a novel approach for detecting heterogeneity in the parameters of IRTree models across subgroups that engage in different response behavior. The approach uses score-based tests to reveal violations of parameter heterogeneity along extraneous person covariates, and it can be employed as a model-based partitioning algorithm to identify sources of differences in the strength of trait-based responding or other response processes. Simulation studies demonstrate generally accurate Type I error rates and sufficient power for metric, ordinal, and categorical person covariates and for different types of test statistics, with the potential to differentiate between different types of parameter heterogeneity. An empirical application illustrates the use of score-based partitioning in the analysis of latent response processes with real data.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142569999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
An alternative closed-form expression for the marginal joint probability distribution of item scores under the random effects generalized partial credit model is presented. The closed-form expression involves a cumulant generating function and is therefore subjected to convexity constraints. As a consequence, complicated moment inequalities are taken into account in maximum likelihood estimation of the parameters of the model, so that the estimation solution is always proper. Another important favorable consequence is that the likelihood function has a single local extreme point, the global maximum. Furthermore, attention is paid to expected a posteriori person parameter estimation, generalizations of the model, and testing the goodness-of-fit of the model. Procedures proposed are demonstrated in an illustrative example.
{"title":"A convexity-constrained parameterization of the random effects generalized partial credit model.","authors":"David J Hessen","doi":"10.1111/bmsp.12365","DOIUrl":"https://doi.org/10.1111/bmsp.12365","url":null,"abstract":"<p><p>An alternative closed-form expression for the marginal joint probability distribution of item scores under the random effects generalized partial credit model is presented. The closed-form expression involves a cumulant generating function and is therefore subjected to convexity constraints. As a consequence, complicated moment inequalities are taken into account in maximum likelihood estimation of the parameters of the model, so that the estimation solution is always proper. Another important favorable consequence is that the likelihood function has a single local extreme point, the global maximum. Furthermore, attention is paid to expected a posteriori person parameter estimation, generalizations of the model, and testing the goodness-of-fit of the model. Procedures proposed are demonstrated in an illustrative example.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142513386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Recently Variational Autoencoders (VAEs) have been proposed as a method to estimate high dimensional Item Response Theory (IRT) models on large datasets. Although these improve the efficiency of estimation drastically compared to traditional methods, they have no natural way to deal with missing values. In this paper, we adapt three existing methods from the VAE literature to the IRT setting and propose one new method. We compare the performance of the different VAE-based methods to each other and to marginal maximum likelihood estimation for increasing levels of missing data in a simulation study for both three- and ten-dimensional IRT models. Additionally, we demonstrate the use of the VAE-based models on an existing algebra test dataset. Results confirm that VAE-based methods are a time-efficient alternative to marginal maximum likelihood, but that a larger number of importance-weighted samples are needed when the proportion of missing values is large.
{"title":"Handling missing data in variational autoencoder based item response theory.","authors":"Karel Veldkamp, Raoul Grasman, Dylan Molenaar","doi":"10.1111/bmsp.12363","DOIUrl":"https://doi.org/10.1111/bmsp.12363","url":null,"abstract":"<p><p>Recently Variational Autoencoders (VAEs) have been proposed as a method to estimate high dimensional Item Response Theory (IRT) models on large datasets. Although these improve the efficiency of estimation drastically compared to traditional methods, they have no natural way to deal with missing values. In this paper, we adapt three existing methods from the VAE literature to the IRT setting and propose one new method. We compare the performance of the different VAE-based methods to each other and to marginal maximum likelihood estimation for increasing levels of missing data in a simulation study for both three- and ten-dimensional IRT models. Additionally, we demonstrate the use of the VAE-based models on an existing algebra test dataset. Results confirm that VAE-based methods are a time-efficient alternative to marginal maximum likelihood, but that a larger number of importance-weighted samples are needed when the proportion of missing values is large.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-10-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142513387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We consider the problem of determining the maximum value of the point-polyserial correlation between a random variable with an assigned continuous distribution and an ordinal random variable with categories, which are assigned the first natural values , and arbitrary probabilities . For different parametric distributions, we derive a closed-form formula for the maximal point-polyserial correlation as a function of the and of the distribution's parameters; we devise an algorithm for obtaining its maximum value numerically for any given . These maximum values and the features of the corresponding -point discrete random variables are discussed with respect to the underlying continuous distribution. Furthermore, we prove that if we do not assign the values of the ordinal random variable a priori but instead include them in the optimization problem, this latter approach is equivalent to the optimal quantization problem. In some circumstances, it leads to a significant increase in the maximum value of the point-polyserial correlation. An application to real data exemplifies the main findings. A comparison between the discretization leading to the maximum point-polyserial correlation and those obtained from optimal quantization and moment matching is sketched.
我们考虑的问题是确定一个具有指定连续分布的随机变量与一个具有 k $$ k $$ 类别的序数随机变量之间的点-序列相关性的最大值,这些类别被赋予前 k $$ k $$ 个自然值 1 , 2 , ... , k $$ 1,2,dots, k $$ 以及任意概率 p i $$ {p}_i $$。对于不同的参数分布,我们推导出了最大点-多序列相关性的闭式公式,它是 p i $$ {p}_i $$ 和分布参数的函数;我们还设计了一种算法,用于在任何给定 k $$ k $$ 的情况下数值求取其最大值。我们讨论了这些最大值以及相应 k $$ k $$ 点离散型随机变量与基本连续分布的关系。此外,我们还证明,如果我们不先验地分配顺序随机变量的值,而是将它们纳入优化问题,那么后一种方法就等同于最优量化问题。在某些情况下,它能显著提高点-多序列相关性的最大值。对真实数据的应用举例说明了主要发现。我们还将对获得最大点-多序列相关性的离散化方法与最优量化和矩匹配方法进行比较。
{"title":"Maximal point-polyserial correlation for non-normal random distributions.","authors":"Alessandro Barbiero","doi":"10.1111/bmsp.12362","DOIUrl":"10.1111/bmsp.12362","url":null,"abstract":"<p><p>We consider the problem of determining the maximum value of the point-polyserial correlation between a random variable with an assigned continuous distribution and an ordinal random variable with <math> <semantics> <mrow><mrow><mi>k</mi></mrow> </mrow> <annotation>$$ k $$</annotation></semantics> </math> categories, which are assigned the first <math> <semantics> <mrow><mrow><mi>k</mi></mrow> </mrow> <annotation>$$ k $$</annotation></semantics> </math> natural values <math> <semantics> <mrow><mrow><mn>1</mn> <mo>,</mo> <mn>2</mn> <mo>,</mo> <mtext>…</mtext> <mo>,</mo> <mi>k</mi></mrow> </mrow> <annotation>$$ 1,2,dots, k $$</annotation></semantics> </math> , and arbitrary probabilities <math> <semantics> <mrow> <mrow> <msub><mrow><mi>p</mi></mrow> <mrow><mi>i</mi></mrow> </msub> </mrow> </mrow> <annotation>$$ {p}_i $$</annotation></semantics> </math> . For different parametric distributions, we derive a closed-form formula for the maximal point-polyserial correlation as a function of the <math> <semantics> <mrow> <mrow> <msub><mrow><mi>p</mi></mrow> <mrow><mi>i</mi></mrow> </msub> </mrow> </mrow> <annotation>$$ {p}_i $$</annotation></semantics> </math> and of the distribution's parameters; we devise an algorithm for obtaining its maximum value numerically for any given <math> <semantics> <mrow><mrow><mi>k</mi></mrow> </mrow> <annotation>$$ k $$</annotation></semantics> </math> . These maximum values and the features of the corresponding <math> <semantics> <mrow><mrow><mi>k</mi></mrow> </mrow> <annotation>$$ k $$</annotation></semantics> </math> -point discrete random variables are discussed with respect to the underlying continuous distribution. Furthermore, we prove that if we do not assign the values of the ordinal random variable a priori but instead include them in the optimization problem, this latter approach is equivalent to the optimal quantization problem. In some circumstances, it leads to a significant increase in the maximum value of the point-polyserial correlation. An application to real data exemplifies the main findings. A comparison between the discretization leading to the maximum point-polyserial correlation and those obtained from optimal quantization and moment matching is sketched.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142481466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Lourens Waldorp, Jolanda Kossakowski, Han L J van der Maas
Networks (graphs) in psychology are often restricted to settings without interventions. Here we consider a framework borrowed from biology that involves multiple interventions from different contexts (observations and experiments) in a single analysis. The method is called perturbation graphs. In gene regulatory networks, the induced change in one gene is measured on all other genes in the analysis, thereby assessing possible causal relations. This is repeated for each gene in the analysis. A perturbation graph leads to the correct set of causes (not nec-essarily direct causes). Subsequent pruning of paths in the graph (called transitive reduction) should reveal direct causes. We show that transitive reduction will not in general lead to the correct underlying graph. We also show that invariant causal prediction is a generalisation of the perturbation graph method and does reveal direct causes, thereby replacing transitive re-duction. We conclude that perturbation graphs provide a promising new tool for experimental designs in psychology, and combined with invariant causal prediction make it possible to re-veal direct causes instead of causal paths. As an illustration we apply these ideas to a data set about attitudes on meat consumption and to a time series of a patient diagnosed with major depression disorder.
{"title":"Perturbation graphs, invariant causal prediction and causal relations in psychology.","authors":"Lourens Waldorp, Jolanda Kossakowski, Han L J van der Maas","doi":"10.1111/bmsp.12361","DOIUrl":"https://doi.org/10.1111/bmsp.12361","url":null,"abstract":"<p><p>Networks (graphs) in psychology are often restricted to settings without interventions. Here we consider a framework borrowed from biology that involves multiple interventions from different contexts (observations and experiments) in a single analysis. The method is called perturbation graphs. In gene regulatory networks, the induced change in one gene is measured on all other genes in the analysis, thereby assessing possible causal relations. This is repeated for each gene in the analysis. A perturbation graph leads to the correct set of causes (not nec-essarily direct causes). Subsequent pruning of paths in the graph (called transitive reduction) should reveal direct causes. We show that transitive reduction will not in general lead to the correct underlying graph. We also show that invariant causal prediction is a generalisation of the perturbation graph method and does reveal direct causes, thereby replacing transitive re-duction. We conclude that perturbation graphs provide a promising new tool for experimental designs in psychology, and combined with invariant causal prediction make it possible to re-veal direct causes instead of causal paths. As an illustration we apply these ideas to a data set about attitudes on meat consumption and to a time series of a patient diagnosed with major depression disorder.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142481470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}