{"title":"Asymptotic results for nonparametric regression estimators after sufficient dimension reduction estimation","authors":"Liliana Forzani, Daniela Rodriguez, Mariela Sued","doi":"10.1007/s11749-024-00932-y","DOIUrl":null,"url":null,"abstract":"<p>Prediction, in regression and classification, is one of the main aims in modern data science. When the number of predictors is large, a common first step is to reduce the dimension of the data. Sufficient dimension reduction (SDR) is a well-established paradigm of reduction that keeps all the relevant information in the covariates <i>X</i> that is necessary for the prediction of <i>Y</i>. In practice, SDR has been successfully used as an exploratory tool for modeling after estimation of the sufficient reduction. Nevertheless, even if the estimated reduction is a consistent estimator of the population, there is no theory supporting this step when nonparametric regression is used in the imputed estimator. In this paper, we show that the asymptotic distribution of the nonparametric regression estimator remains unchanged whether the true SDR or its estimator is used. This result allows making inferences, for example, computing confidence intervals for the regression function, thereby avoiding the curse of dimensionality.</p>","PeriodicalId":51189,"journal":{"name":"Test","volume":"70 1","pages":""},"PeriodicalIF":1.2000,"publicationDate":"2024-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Test","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1007/s11749-024-00932-y","RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"STATISTICS & PROBABILITY","Score":null,"Total":0}
引用次数: 0
Abstract
Prediction, in regression and classification, is one of the main aims in modern data science. When the number of predictors is large, a common first step is to reduce the dimension of the data. Sufficient dimension reduction (SDR) is a well-established paradigm of reduction that keeps all the relevant information in the covariates X that is necessary for the prediction of Y. In practice, SDR has been successfully used as an exploratory tool for modeling after estimation of the sufficient reduction. Nevertheless, even if the estimated reduction is a consistent estimator of the population, there is no theory supporting this step when nonparametric regression is used in the imputed estimator. In this paper, we show that the asymptotic distribution of the nonparametric regression estimator remains unchanged whether the true SDR or its estimator is used. This result allows making inferences, for example, computing confidence intervals for the regression function, thereby avoiding the curse of dimensionality.
回归和分类中的预测是现代数据科学的主要目标之一。当预测因子数量较多时,第一步通常是降低数据维度。充分降维(SDR)是一种行之有效的降维范式,它能保留协变量 X 中对预测 Y 必不可少的所有相关信息。然而,即使估计出的还原是一个一致的总体估计器,但当在估算估计器中使用非参数回归时,并没有理论支持这一步骤。在本文中,我们证明了无论使用真实 SDR 还是其估计值,非参数回归估计值的渐近分布都保持不变。这一结果允许进行推论,例如计算回归函数的置信区间,从而避免了维度诅咒。
期刊介绍:
TEST is an international journal of Statistics and Probability, sponsored by the Spanish Society of Statistics and Operations Research. English is the official language of the journal.
The emphasis of TEST is placed on papers containing original theoretical contributions of direct or potential value in applications. In this respect, the methodological contents are considered to be crucial for the papers published in TEST, but the practical implications of the methodological aspects are also relevant. Original sound manuscripts on either well-established or emerging areas in the scope of the journal are welcome.
One volume is published annually in four issues. In addition to the regular contributions, each issue of TEST contains an invited paper from a world-wide recognized outstanding statistician on an up-to-date challenging topic, including discussions.