{"title":"PRINCIPAL COMPONENTS AND THE PROBLEM OF MULTICOLLINEARITY","authors":"B. Morzuch","doi":"10.1017/S0163548400002478","DOIUrl":null,"url":null,"abstract":"Multicollinearity among independent variables within a regress ion model is one of the most frequently encountered problems faced by the applied researcher. In a recent article in this Journal (Willis, e1 a/.) , a catalog of\"remedies\" for multicollinearity was presented to assist in reducing its level and associated adverse consequences . One of these remediesprincipal componentswas suggested as a method oftransforming a set of collinear explanatory variables into new variables that are orthogonal to each other with the first few of these transformed va riables accounting for the majority of the variability in the origina l data set. In principal components regression , a transformed variable is determined to be important a nd included or unimportant and excluded in the regression model depending upon the size of the characteristic root (eigenvalue) associated with its corresponding characteristic vector (eigenvector) (Massy), the statistical significance of its regression coefficient (Mittelhammer and Baritelle) , or some combination of eigenvalue size and correlation with the dependent variable (Johnson, et a/.) . Unfortunately, this technique is widely abused and misunderstood by the applied researcher. Confusion exists with respect to (I) its relationship to other orthogonalization techniques; (2) the meaning of the orthogonalized components and the necessity of transforming the principal component estimators back to the original parameter space; (3) the implications of deleting components and the correspondence of this technique to a particular type of restricted least squares estimator; (4) the proper way to delete components and evaluate these implied restrictions; and (5) actual implementation of this estimation procedure via available computer routines . The purpose of this note, therefore, is to place the technique of principal components in perspective and to suggest a methodology for implementing this technique correctly.","PeriodicalId":421915,"journal":{"name":"Journal of the Northeastern Agricultural Economics Council","volume":"44 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1980-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of the Northeastern Agricultural Economics Council","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1017/S0163548400002478","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 10
Abstract
Multicollinearity among independent variables within a regress ion model is one of the most frequently encountered problems faced by the applied researcher. In a recent article in this Journal (Willis, e1 a/.) , a catalog of"remedies" for multicollinearity was presented to assist in reducing its level and associated adverse consequences . One of these remediesprincipal componentswas suggested as a method oftransforming a set of collinear explanatory variables into new variables that are orthogonal to each other with the first few of these transformed va riables accounting for the majority of the variability in the origina l data set. In principal components regression , a transformed variable is determined to be important a nd included or unimportant and excluded in the regression model depending upon the size of the characteristic root (eigenvalue) associated with its corresponding characteristic vector (eigenvector) (Massy), the statistical significance of its regression coefficient (Mittelhammer and Baritelle) , or some combination of eigenvalue size and correlation with the dependent variable (Johnson, et a/.) . Unfortunately, this technique is widely abused and misunderstood by the applied researcher. Confusion exists with respect to (I) its relationship to other orthogonalization techniques; (2) the meaning of the orthogonalized components and the necessity of transforming the principal component estimators back to the original parameter space; (3) the implications of deleting components and the correspondence of this technique to a particular type of restricted least squares estimator; (4) the proper way to delete components and evaluate these implied restrictions; and (5) actual implementation of this estimation procedure via available computer routines . The purpose of this note, therefore, is to place the technique of principal components in perspective and to suggest a methodology for implementing this technique correctly.
回归模型中自变量之间的多重共线性是应用研究者最常遇到的问题之一。在该杂志最近的一篇文章中(Willis, e1 a/.),提出了多重共线性的“补救措施”目录,以帮助降低其水平和相关的不良后果。其中一种补救措施-主成分-被建议作为将一组共线性解释变量转换为相互正交的新变量的方法,这些转换后的变量中的前几个占原始数据集中的大部分可变性。在主成分回归中,根据与其对应的特征向量(特征向量)(Massy)相关联的特征根(特征值)的大小,其回归系数(Mittelhammer和Baritelle)的统计显著性,或特征值大小和与因变量的相关性的某种组合(Johnson, et a/.),确定转换后的变量是重要的,包括或不重要并排除在回归模型中。不幸的是,这种技术被应用研究人员广泛滥用和误解。关于(1)它与其他正交化技术的关系存在混淆;(2)正交化分量的含义和将主分量估计量变换回原参数空间的必要性;(3)删除分量的含义以及该技术与特定类型的受限最小二乘估计的对应关系;(4)删除组件和评估这些隐含限制的正确方法;(5)通过可用的计算机例程实际实现该估计过程。因此,本文的目的是对主成分技术进行透视,并提出正确实现该技术的方法。