Pub Date : 2024-05-08DOI: 10.1007/s10182-024-00502-5
Alexander Gerharz, Andreas Groll, Gunther Schauberger
In this article, a new kind of interpretable machine learning method is presented, which can help to understand the partition of the feature space into predicted classes in a classification model using quantile shifts, and this way make the underlying statistical or machine learning model more trustworthy. Basically, real data points (or specific points of interest) are used and the changes of the prediction after slightly raising or decreasing specific features are observed. By comparing the predictions before and after the shifts, under certain conditions the observed changes in the predictions can be interpreted as neighborhoods of the classes with regard to the shifted features. Chord diagrams are used to visualize the observed changes. For illustration, this quantile shift method (QSM) is applied to an artificial example with medical labels and a real data example.
{"title":"Deducing neighborhoods of classes from a fitted model","authors":"Alexander Gerharz, Andreas Groll, Gunther Schauberger","doi":"10.1007/s10182-024-00502-5","DOIUrl":"10.1007/s10182-024-00502-5","url":null,"abstract":"<div><p>In this article, a new kind of interpretable machine learning method is presented, which can help to understand the partition of the feature space into predicted classes in a classification model using quantile shifts, and this way make the underlying statistical or machine learning model more trustworthy. Basically, real data points (or specific points of interest) are used and the changes of the prediction after slightly raising or decreasing specific features are observed. By comparing the predictions before and after the shifts, under certain conditions the observed changes in the predictions can be interpreted as neighborhoods of the classes with regard to the shifted features. Chord diagrams are used to visualize the observed changes. For illustration, this quantile shift method (QSM) is applied to an artificial example with medical labels and a real data example.</p></div>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2024-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10182-024-00502-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140936567","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-13DOI: 10.1007/s10182-024-00498-y
Francesca Di Iorio, Riccardo Lucchetti, Rosaria Simone
In this paper, we propose a portmanteau test for misspecification in combination of uniform and binomial (CUB) models for the analysis of ordered rating data. Specifically, the test we build belongs to the class of information matrix (IM) tests that are based on the information matrix equality. Monte Carlo evidence indicates that the test has excellent properties in finite samples in terms of actual size and power versus several alternatives. Differently from other tests of the IM family, finite-sample adjustments based on the bootstrap seem to be unnecessary. An empirical application is also provided to illustrate how the IM test can be used to supplement model validation and selection.
在本文中,我们提出了一种用于分析有序评级数据的统一和二项(CUB)组合模型的波特曼检验法(portmanteau test)。具体来说,我们建立的检验属于基于信息矩阵相等的信息矩阵(IM)检验。蒙特卡洛证据表明,在有限样本中,该检验在实际规模和功率方面相对于几种备选方案都具有出色的特性。与 IM 系列的其他检验不同,基于引导的有限样本调整似乎是不必要的。本文还提供了一个经验应用,以说明如何使用 IM 检验来补充模型验证和选择。
{"title":"Testing distributional assumptions in CUB models for the analysis of rating data","authors":"Francesca Di Iorio, Riccardo Lucchetti, Rosaria Simone","doi":"10.1007/s10182-024-00498-y","DOIUrl":"10.1007/s10182-024-00498-y","url":null,"abstract":"<div><p>In this paper, we propose a <i>portmanteau</i> test for misspecification in combination of uniform and binomial (CUB) models for the analysis of ordered rating data. Specifically, the test we build belongs to the class of information matrix (IM) tests that are based on the information matrix equality. Monte Carlo evidence indicates that the test has excellent properties in finite samples in terms of actual size and power versus several alternatives. Differently from other tests of the IM family, finite-sample adjustments based on the bootstrap seem to be unnecessary. An empirical application is also provided to illustrate how the IM test can be used to supplement model validation and selection.</p></div>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2024-04-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10182-024-00498-y.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140582934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-12DOI: 10.1007/s10182-024-00499-x
Jan Beran, Jeremy Näscher, Fabian Pietsch, Stephan Walterspacher
A frequent problem in applied time series analysis is the identification of dominating periodic components. A particularly difficult task is to distinguish deterministic periodic signals from periodic long memory. In this paper, a family of test statistics based on Whittle’s Gaussian log-likelihood approximation is proposed. Asymptotic critical regions and bounds for the asymptotic power are derived. In cases where a deterministic periodic signal and periodic long memory share the same frequency, consistency and rates of type II error probabilities depend on the long-memory parameter. Simulations and an application to respiratory muscle training data illustrate the results.
在应用时间序列分析中,一个经常遇到的问题是如何识别占主导地位的周期成分。一个特别困难的任务是将确定性周期信号与周期性长记忆区分开来。本文提出了基于惠特尔高斯对数似然近似的检验统计量系列。推导出了渐近临界区和渐近功率的边界。在确定性周期信号和周期性长记忆共享相同频率的情况下,一致性和 II 型错误概率率取决于长记忆参数。模拟和呼吸肌训练数据的应用说明了这些结果。
{"title":"Testing for periodicity at an unknown frequency under cyclic long memory, with applications to respiratory muscle training","authors":"Jan Beran, Jeremy Näscher, Fabian Pietsch, Stephan Walterspacher","doi":"10.1007/s10182-024-00499-x","DOIUrl":"https://doi.org/10.1007/s10182-024-00499-x","url":null,"abstract":"<p>A frequent problem in applied time series analysis is the identification of dominating periodic components. A particularly difficult task is to distinguish deterministic periodic signals from periodic long memory. In this paper, a family of test statistics based on Whittle’s Gaussian log-likelihood approximation is proposed. Asymptotic critical regions and bounds for the asymptotic power are derived. In cases where a deterministic periodic signal and periodic long memory share the same frequency, consistency and rates of type II error probabilities depend on the long-memory parameter. Simulations and an application to respiratory muscle training data illustrate the results.</p>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2024-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140582923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-03DOI: 10.1007/s10182-024-00497-z
Oliver Dürr, Stefan Hörtling, Danil Dold, Ivonne Kovylov, Beate Sick
Black-box variational inference (BBVI) is a technique to approximate the posterior of Bayesian models by optimization. Similar to MCMC, the user only needs to specify the model; then, the inference procedure is done automatically. In contrast to MCMC, BBVI scales to many observations, is faster for some applications, and can take advantage of highly optimized deep learning frameworks since it can be formulated as a minimization task. In the case of complex posteriors, however, other state-of-the-art BBVI approaches often yield unsatisfactory posterior approximations. This paper presents Bernstein flow variational inference (BF-VI), a robust and easy-to-use method flexible enough to approximate complex multivariate posteriors. BF-VI combines ideas from normalizing flows and Bernstein polynomial-based transformation models. In benchmark experiments, we compare BF-VI solutions with exact posteriors, MCMC solutions, and state-of-the-art BBVI methods, including normalizing flow-based BBVI. We show for low-dimensional models that BF-VI accurately approximates the true posterior; in higher-dimensional models, BF-VI compares favorably against other BBVI methods. Further, using BF-VI, we develop a Bayesian model for the semi-structured melanoma challenge data, combining a CNN model part for image data with an interpretable model part for tabular data, and demonstrate, for the first time, the use of BBVI in semi-structured models.
{"title":"Bernstein flows for flexible posteriors in variational Bayes","authors":"Oliver Dürr, Stefan Hörtling, Danil Dold, Ivonne Kovylov, Beate Sick","doi":"10.1007/s10182-024-00497-z","DOIUrl":"10.1007/s10182-024-00497-z","url":null,"abstract":"<div><p>Black-box variational inference (BBVI) is a technique to approximate the posterior of Bayesian models by optimization. Similar to MCMC, the user only needs to specify the model; then, the inference procedure is done automatically. In contrast to MCMC, BBVI scales to many observations, is faster for some applications, and can take advantage of highly optimized deep learning frameworks since it can be formulated as a minimization task. In the case of complex posteriors, however, other state-of-the-art BBVI approaches often yield unsatisfactory posterior approximations. This paper presents Bernstein flow variational inference (BF-VI), a robust and easy-to-use method flexible enough to approximate complex multivariate posteriors. BF-VI combines ideas from normalizing flows and Bernstein polynomial-based transformation models. In benchmark experiments, we compare BF-VI solutions with exact posteriors, MCMC solutions, and state-of-the-art BBVI methods, including normalizing flow-based BBVI. We show for low-dimensional models that BF-VI accurately approximates the true posterior; in higher-dimensional models, BF-VI compares favorably against other BBVI methods. Further, using BF-VI, we develop a Bayesian model for the semi-structured melanoma challenge data, combining a CNN model part for image data with an interpretable model part for tabular data, and demonstrate, for the first time, the use of BBVI in semi-structured models.</p></div>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2024-04-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10182-024-00497-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140582930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-03DOI: 10.1007/s10182-024-00492-4
Jens Lichter, Paul F V Wiemann, Thomas Kneib
Markov chain Monte Carlo (MCMC)-based simulation approaches are by far the most common method in Bayesian inference to access the posterior distribution. Recently, motivated by successes in machine learning, variational inference (VI) has gained in interest in statistics since it promises a computationally efficient alternative to MCMC enabling approximate access to the posterior. Classical approaches such as mean-field VI (MFVI), however, are based on the strong mean-field assumption for the approximate posterior where parameters or parameter blocks are assumed to be mutually independent. As a consequence, parameter uncertainties are often underestimated and alternatives such as semi-implicit VI (SIVI) have been suggested to avoid the mean-field assumption and to improve uncertainty estimates. SIVI uses a hierarchical construction of the variational parameters to restore parameter dependencies and relies on a highly flexible implicit mixing distribution whose probability density function is not analytic but samples can be taken via a stochastic procedure. With this paper, we investigate how different forms of VI perform in semiparametric additive regression models as one of the most important fields of application of Bayesian inference in statistics. A particular focus is on the ability of the rivalling approaches to quantify uncertainty, especially with correlated covariates that are likely to aggravate the difficulties of simplifying VI assumptions. Moreover, we propose a method, where we combine both advantages of MFVI and SIVI and compare its performance. The different VI approaches are studied in comparison with MCMC in simulations and an application to tree height models of douglas fir based on a large-scale forestry data set.
基于马尔科夫链蒙特卡罗(MCMC)的模拟方法是贝叶斯推理中迄今为止最常用的获取后验分布的方法。最近,在机器学习取得成功的推动下,变分推理(VI)在统计学中越来越受到关注,因为它有望成为 MCMC 的一种计算高效的替代方法,能够近似访问后验分布。然而,均值场变分推理(MFVI)等经典方法是基于近似后验的强均值场假设,其中参数或参数块被假定为相互独立的。因此,参数的不确定性往往被低估,人们提出了半隐式 VI(SIVI)等替代方法,以避免均值场假设并改进不确定性估计。SIVI 使用变分参数的分层结构来恢复参数依赖关系,并依赖于高度灵活的隐式混合分布,其概率密度函数不是解析的,但可以通过随机过程取样。本文研究了不同形式的 VI 在半参数加法回归模型中的表现,该模型是贝叶斯推理在统计学中最重要的应用领域之一。本文特别关注了不同方法量化不确定性的能力,尤其是在相关协变量可能加剧简化 VI 假设困难的情况下。此外,我们还提出了一种方法,该方法结合了 MFVI 和 SIVI 的优点,并对其性能进行了比较。我们将不同的 VI 方法与模拟 MCMC 进行了比较研究,并将其应用于基于大规模林业数据集的道格拉斯杉树高模型。
{"title":"Variational inference: uncertainty quantification in additive models","authors":"Jens Lichter, Paul F V Wiemann, Thomas Kneib","doi":"10.1007/s10182-024-00492-4","DOIUrl":"10.1007/s10182-024-00492-4","url":null,"abstract":"<div><p>Markov chain Monte Carlo (MCMC)-based simulation approaches are by far the most common method in Bayesian inference to access the posterior distribution. Recently, motivated by successes in machine learning, variational inference (VI) has gained in interest in statistics since it promises a computationally efficient alternative to MCMC enabling approximate access to the posterior. Classical approaches such as mean-field VI (MFVI), however, are based on the strong mean-field assumption for the approximate posterior where parameters or parameter blocks are assumed to be mutually independent. As a consequence, parameter uncertainties are often underestimated and alternatives such as semi-implicit VI (SIVI) have been suggested to avoid the mean-field assumption and to improve uncertainty estimates. SIVI uses a hierarchical construction of the variational parameters to restore parameter dependencies and relies on a highly flexible implicit mixing distribution whose probability density function is not analytic but samples can be taken via a stochastic procedure. With this paper, we investigate how different forms of VI perform in semiparametric additive regression models as one of the most important fields of application of Bayesian inference in statistics. A particular focus is on the ability of the rivalling approaches to quantify uncertainty, especially with correlated covariates that are likely to aggravate the difficulties of simplifying VI assumptions. Moreover, we propose a method, where we combine both advantages of MFVI and SIVI and compare its performance. The different VI approaches are studied in comparison with MCMC in simulations and an application to tree height models of douglas fir based on a large-scale forestry data set.</p></div>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2024-04-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10182-024-00492-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140582932","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-01DOI: 10.1007/s10182-024-00496-0
Cristina O. Chavez-Chong, Cécile Hardouin, Ana-Karina Fermin
This work proposes a new method for building an explanatory spatial autoregressive model in a multicollinearity context. We use Ridge regularization to bypass the collinearity issue. We present new estimation algorithms that allow for the estimation of the regression coefficients as well as the spatial dependence parameter. A spatial cross-validation procedure is used to tune the regularization parameter. In fact, ordinary cross-validation techniques are not applicable to spatially dependent observations. Variable importance is assessed by permutation tests since classical tests are not valid after Ridge regularization. We assess the performance of our methodology through numerical experiments conducted on simulated synthetic data. Finally, we apply our method to a real data set and evaluate the impact of some socioeconomic variables on the COVID-19 intensity in France.
{"title":"Ridge regularization for spatial autoregressive models with multicollinearity issues","authors":"Cristina O. Chavez-Chong, Cécile Hardouin, Ana-Karina Fermin","doi":"10.1007/s10182-024-00496-0","DOIUrl":"https://doi.org/10.1007/s10182-024-00496-0","url":null,"abstract":"<p>This work proposes a new method for building an explanatory spatial autoregressive model in a multicollinearity context. We use Ridge regularization to bypass the collinearity issue. We present new estimation algorithms that allow for the estimation of the regression coefficients as well as the spatial dependence parameter. A spatial cross-validation procedure is used to tune the regularization parameter. In fact, ordinary cross-validation techniques are not applicable to spatially dependent observations. Variable importance is assessed by permutation tests since classical tests are not valid after Ridge regularization. We assess the performance of our methodology through numerical experiments conducted on simulated synthetic data. Finally, we apply our method to a real data set and evaluate the impact of some socioeconomic variables on the COVID-19 intensity in France.</p>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140582922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-14DOI: 10.1007/s10182-024-00495-1
Philip Buczak, Andreas Groll, Markus Pauly, Jakob Rehof, Daniel Horn
Hyperparameter tuning is one of the most time-consuming parts in machine learning. Despite the existence of modern optimization algorithms that minimize the number of evaluations needed, evaluations of a single setting may still be expensive. Usually a resampling technique is used, where the machine learning method has to be fitted a fixed number of k times on different training datasets. The respective mean performance of the k fits is then used as performance estimator. Many hyperparameter settings could be discarded after less than k resampling iterations if they are clearly inferior to high-performing settings. However, resampling is often performed until the very end, wasting a lot of computational effort. To this end, we propose the sequential random search (SQRS) which extends the regular random search algorithm by a sequential testing procedure aimed at detecting and eliminating inferior parameter configurations early. We compared our SQRS with regular random search using multiple publicly available regression and classification datasets. Our simulation study showed that the SQRS is able to find similarly well-performing parameter settings while requiring noticeably fewer evaluations. Our results underscore the potential for integrating sequential tests into hyperparameter tuning.
超参数调整是机器学习中最耗时的部分之一。尽管现代优化算法可以最大限度地减少所需的评估次数,但对单个设置的评估仍可能非常昂贵。通常会使用重采样技术,即在不同的训练数据集上对机器学习方法进行固定次数的 k 次拟合。然后将 k 次拟合各自的平均性能作为性能估计值。如果许多超参数设置明显不如高性能设置,那么可以在少于 k 次的重采样迭代后将其舍弃。然而,重采样往往要到最后才进行,浪费了大量的计算资源。为此,我们提出了顺序随机搜索(SQRS),它通过一个顺序测试程序扩展了常规随机搜索算法,旨在及早检测和消除劣质参数配置。我们使用多个公开的回归和分类数据集对 SQRS 和常规随机搜索进行了比较。我们的模拟研究表明,SQRS 能够找到类似的性能良好的参数设置,而所需的评估次数却明显减少。我们的结果强调了将顺序测试整合到超参数调整中的潜力。
{"title":"Using sequential statistical tests for efficient hyperparameter tuning","authors":"Philip Buczak, Andreas Groll, Markus Pauly, Jakob Rehof, Daniel Horn","doi":"10.1007/s10182-024-00495-1","DOIUrl":"10.1007/s10182-024-00495-1","url":null,"abstract":"<div><p>Hyperparameter tuning is one of the most time-consuming parts in machine learning. Despite the existence of modern optimization algorithms that minimize the number of evaluations needed, evaluations of a single setting may still be expensive. Usually a resampling technique is used, where the machine learning method has to be fitted a fixed number of <i>k</i> times on different training datasets. The respective mean performance of the <i>k</i> fits is then used as performance estimator. Many hyperparameter settings could be discarded after less than <i>k</i> resampling iterations if they are clearly inferior to high-performing settings. However, resampling is often performed until the very end, wasting a lot of computational effort. To this end, we propose the sequential random search (SQRS) which extends the regular random search algorithm by a sequential testing procedure aimed at detecting and eliminating inferior parameter configurations early. We compared our SQRS with regular random search using multiple publicly available regression and classification datasets. Our simulation study showed that the SQRS is able to find similarly well-performing parameter settings while requiring noticeably fewer evaluations. Our results underscore the potential for integrating sequential tests into hyperparameter tuning.</p></div>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2024-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10182-024-00495-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140124518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-11DOI: 10.1007/s10182-024-00494-2
Claudio Agostinelli, Luca Greco, Giovanni Saraceno
We consider, robust estimation of wrapped models to multivariate circular data that are points on the surface of a p-torus based on the weighted likelihood methodology. Robust model fitting is achieved by a set of weighted likelihood estimating equations, based on the computation of data dependent weights aimed to down-weight anomalous values, such as unexpected directions that do not share the main pattern of the bulk of the data. Weighted likelihood estimating equations with weights evaluated on the torus or obtained after unwrapping the data onto the Euclidean space are proposed and compared. Asymptotic properties and robustness features of the estimators under study have been studied, whereas their finite sample behavior has been investigated by Monte Carlo numerical experiment and real data examples.
{"title":"Weighted likelihood methods for robust fitting of wrapped models for p-torus data","authors":"Claudio Agostinelli, Luca Greco, Giovanni Saraceno","doi":"10.1007/s10182-024-00494-2","DOIUrl":"https://doi.org/10.1007/s10182-024-00494-2","url":null,"abstract":"<p>We consider, robust estimation of wrapped models to multivariate circular data that are points on the surface of a <i>p</i>-torus based on the weighted likelihood methodology. Robust model fitting is achieved by a set of weighted likelihood estimating equations, based on the computation of data dependent weights aimed to down-weight anomalous values, such as unexpected directions that do not share the main pattern of the bulk of the data. Weighted likelihood estimating equations with weights evaluated on the torus or obtained after unwrapping the data onto the Euclidean space are proposed and compared. Asymptotic properties and robustness features of the estimators under study have been studied, whereas their finite sample behavior has been investigated by Monte Carlo numerical experiment and real data examples.</p>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2024-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140116179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-06DOI: 10.1007/s10182-024-00493-3
Abstract
In small area estimation, the sample size is so small that direct estimators have seldom enough adequate precision. Therefore, it is common to use auxiliary data via covariates and produce estimators that combine them with direct data. Nevertheless, it is not uncommon for covariates to be measured with error, leading to inconsistent estimators. Area-level models accounting for measurement error (ME) in covariates have been proposed, and they usually assume that the errors are an i.i.d. Gaussian model. However, there might be situations in which this assumption is violated especially when covariates present severe outlying values that cannot be cached by the Gaussian distribution. To overcome this problem, we propose to model the ME through sub-Gaussian (alpha)-stable (SG(alpha)S) distribution, a flexible distribution that accommodates different types of outlying observations and also Gaussian data as a special case when (alpha =2). The SG(alpha)S distribution is a generalization of the Gaussian distribution that allows for skewness and heavy tails by adding an extra parameter, (alpha in (0,2]), to control tail behaviour. The model parameters are estimated in a fully Bayesian framework. The performance of the proposal is illustrated by applying to real data and some simulation studies.
摘要 在小面积估算中,样本量非常小,直接估算器很少有足够的精度。因此,通常通过协变量使用辅助数据,并将其与直接数据结合生成估算器。然而,协变量的测量存在误差,导致估计值不一致的情况并不少见。有人提出了考虑协变量测量误差(ME)的区域级模型,这些模型通常假设误差为 i.i.d. 高斯模型。然而,在某些情况下,这一假设可能会被违反,尤其是当协变量出现严重的离群值,而高斯分布无法将其缓存时。为了克服这个问题,我们建议通过亚高斯稳定分布(SG (alpha) S)对 ME 进行建模,这是一种灵活的分布,可以容纳不同类型的离差观测值,当 (alpha =2)时,高斯数据也是一种特殊情况。SG (alpha) S 分布是高斯分布的广义化,通过增加一个额外参数((0,2])来控制尾部行为,从而允许偏斜和重尾。模型参数在完全贝叶斯框架下进行估计。通过应用真实数据和一些模拟研究,说明了该建议的性能。
{"title":"Robust Bayesian small area estimation using the sub-Gaussian $$alpha$$ -stable distribution for measurement error in covariates","authors":"","doi":"10.1007/s10182-024-00493-3","DOIUrl":"https://doi.org/10.1007/s10182-024-00493-3","url":null,"abstract":"<h3>Abstract</h3> <p>In small area estimation, the sample size is so small that direct estimators have seldom enough adequate precision. Therefore, it is common to use auxiliary data via covariates and produce estimators that combine them with direct data. Nevertheless, it is not uncommon for covariates to be measured with error, leading to inconsistent estimators. Area-level models accounting for measurement error (ME) in covariates have been proposed, and they usually assume that the errors are an i.i.d. Gaussian model. However, there might be situations in which this assumption is violated especially when covariates present severe outlying values that cannot be cached by the Gaussian distribution. To overcome this problem, we propose to model the ME through sub-Gaussian <span> <span>(alpha)</span> </span>-stable (SG<span> <span>(alpha)</span> </span>S) distribution, a flexible distribution that accommodates different types of outlying observations and also Gaussian data as a special case when <span> <span>(alpha =2)</span> </span>. The SG<span> <span>(alpha)</span> </span>S distribution is a generalization of the Gaussian distribution that allows for skewness and heavy tails by adding an extra parameter, <span> <span>(alpha in (0,2])</span> </span>, to control tail behaviour. The model parameters are estimated in a fully Bayesian framework. The performance of the proposal is illustrated by applying to real data and some simulation studies.</p>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140043971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-20DOI: 10.1007/s10182-023-00489-5
Christian Aßmann, Jens Boysen-Hogrefe, Markus Pape
Orthonormality constraints are common in reduced rank models. They imply that matrix-variate parameters are given as orthonormal column vectors. However, these orthonormality restrictions do not provide identification for all parameters. For this setup, we show how the remaining identification issue can be handled in a Bayesian analysis via post-processing the sampling output according to an appropriately specified loss function. This extends the possibilities for Bayesian inference in reduced rank regression models with a part of the parameter space restricted to the Stiefel manifold. Besides inference, we also discuss model selection in terms of posterior predictive assessment. We illustrate the proposed approach with a simulation study and an empirical application.
{"title":"Post-processing for Bayesian analysis of reduced rank regression models with orthonormality restrictions","authors":"Christian Aßmann, Jens Boysen-Hogrefe, Markus Pape","doi":"10.1007/s10182-023-00489-5","DOIUrl":"10.1007/s10182-023-00489-5","url":null,"abstract":"<div><p>Orthonormality constraints are common in reduced rank models. They imply that matrix-variate parameters are given as orthonormal column vectors. However, these orthonormality restrictions do not provide identification for all parameters. For this setup, we show how the remaining identification issue can be handled in a Bayesian analysis via post-processing the sampling output according to an appropriately specified loss function. This extends the possibilities for Bayesian inference in reduced rank regression models with a part of the parameter space restricted to the Stiefel manifold. Besides inference, we also discuss model selection in terms of posterior predictive assessment. We illustrate the proposed approach with a simulation study and an empirical application.</p></div>","PeriodicalId":55446,"journal":{"name":"Asta-Advances in Statistical Analysis","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2023-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10182-023-00489-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138818116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}