{"title":"Optimal Covariance Estimation for Condition Number Loss in the Spiked model","authors":"David Donoho, Behrooz Ghorbani","doi":"10.1016/j.ecosta.2024.04.004","DOIUrl":null,"url":null,"abstract":"Consider estimation of the covariance matrix under relative condition number loss , where is the condition number of matrix , and and are the estimated and theoretical covariance matrices. Recent advances in understanding the so-called for , are used here to derive a nonlinear shrinker which is asymptotically optimal among orthogonally-covariant procedures. These advances apply in an asymptotic setting, where the number of variables is comparable to the number of observations . The form of the optimal nonlinearity depends on the aspect ratio of the data matrix and on the top eigenvalue of . For , even dependence on the top eigenvalue can be avoided. The optimal shrinker has three notable properties. First, when is moderately large, it shrinks even very large eigenvalues substantially, by a factor . Second, even for moderate , certain highly statistically significant eigencomponents will be completely suppressed.Third, when is very large, the optimal covariance estimator can be purely diagonal, despite the top theoretical eigenvalue being large and the empirical eigenvalues being highly statistically significant. This aligns with practitioner experience. Alternatively, certain non-optimal intuitively reasonable procedures can have small worst-case relative regret - the simplest being generalized soft thresholding having threshold at the bulk edge and slope above the bulk. For this has at most a few percent relative regret.","PeriodicalId":54125,"journal":{"name":"Econometrics and Statistics","volume":"10 1","pages":""},"PeriodicalIF":2.0000,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Econometrics and Statistics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1016/j.ecosta.2024.04.004","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ECONOMICS","Score":null,"Total":0}
引用次数: 0
Abstract
Consider estimation of the covariance matrix under relative condition number loss , where is the condition number of matrix , and and are the estimated and theoretical covariance matrices. Recent advances in understanding the so-called for , are used here to derive a nonlinear shrinker which is asymptotically optimal among orthogonally-covariant procedures. These advances apply in an asymptotic setting, where the number of variables is comparable to the number of observations . The form of the optimal nonlinearity depends on the aspect ratio of the data matrix and on the top eigenvalue of . For , even dependence on the top eigenvalue can be avoided. The optimal shrinker has three notable properties. First, when is moderately large, it shrinks even very large eigenvalues substantially, by a factor . Second, even for moderate , certain highly statistically significant eigencomponents will be completely suppressed.Third, when is very large, the optimal covariance estimator can be purely diagonal, despite the top theoretical eigenvalue being large and the empirical eigenvalues being highly statistically significant. This aligns with practitioner experience. Alternatively, certain non-optimal intuitively reasonable procedures can have small worst-case relative regret - the simplest being generalized soft thresholding having threshold at the bulk edge and slope above the bulk. For this has at most a few percent relative regret.
期刊介绍:
Econometrics and Statistics is the official journal of the networks Computational and Financial Econometrics and Computational and Methodological Statistics. It publishes research papers in all aspects of econometrics and statistics and comprises of the two sections Part A: Econometrics and Part B: Statistics.