Pub Date : 2024-06-06DOI: 10.1016/j.jspi.2024.106200
Radoslav Harman, Samuel Rosa
The problem of computing an exact experimental design that is optimal for the least-squares estimation of the parameters of a regression model is considered. We show that this problem can be solved via mixed-integer linear programming (MILP) for a wide class of optimality criteria, including the criteria of A-, I-, G- and MV-optimality. This approach improves upon the current state-of-the-art mathematical programming formulation, which uses mixed-integer second-order cone programming. The key idea underlying the MILP formulation is McCormick relaxation, which critically depends on finite interval bounds for the elements of the covariance matrix of the least-squares estimator corresponding to an optimal exact design. We provide both analytic and algorithmic methods for constructing these bounds. We also demonstrate the unique advantages of the MILP approach, such as the possibility of incorporating multiple design constraints into the optimization problem, including constraints on the variances and covariances of the least-squares estimator.
{"title":"Mixed-integer linear programming for computing optimal experimental designs","authors":"Radoslav Harman, Samuel Rosa","doi":"10.1016/j.jspi.2024.106200","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106200","url":null,"abstract":"<div><p>The problem of computing an exact experimental design that is optimal for the least-squares estimation of the parameters of a regression model is considered. We show that this problem can be solved via mixed-integer linear programming (MILP) for a wide class of optimality criteria, including the criteria of A-, I-, G- and MV-optimality. This approach improves upon the current state-of-the-art mathematical programming formulation, which uses mixed-integer second-order cone programming. The key idea underlying the MILP formulation is McCormick relaxation, which critically depends on finite interval bounds for the elements of the covariance matrix of the least-squares estimator corresponding to an optimal exact design. We provide both analytic and algorithmic methods for constructing these bounds. We also demonstrate the unique advantages of the MILP approach, such as the possibility of incorporating multiple design constraints into the optimization problem, including constraints on the variances and covariances of the least-squares estimator.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"234 ","pages":"Article 106200"},"PeriodicalIF":0.9,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141323229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-05DOI: 10.1016/j.jspi.2024.106197
Lazaros Kanellopoulos, Konstadinos Politis
We study some properties and relations for stochastic orders and aging classes related to the Laplace transform. In particular, we show that the NBU class of distributions is closed under convolution. We also obtain results for the ratio of derivatives of the Laplace transform between two distributions.
{"title":"Some results for stochastic orders and aging properties related to the Laplace transform","authors":"Lazaros Kanellopoulos, Konstadinos Politis","doi":"10.1016/j.jspi.2024.106197","DOIUrl":"10.1016/j.jspi.2024.106197","url":null,"abstract":"<div><p>We study some properties and relations for stochastic orders and aging classes related to the Laplace transform. In particular, we show that the NBU<span><math><msub><mrow></mrow><mrow><mtext>Lt</mtext></mrow></msub></math></span> class of distributions is closed under convolution. We also obtain results for the ratio of derivatives of the Laplace transform between two distributions.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"234 ","pages":"Article 106197"},"PeriodicalIF":0.9,"publicationDate":"2024-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141403038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-05DOI: 10.1016/j.jspi.2024.106188
Michael Kohler , Sophie Langer
Convolutional neural networks (CNNs) trained with cross-entropy loss have proven to be extremely successful in classifying images. In recent years, much work has been done to also improve the theoretical understanding of neural networks. Nevertheless, it seems limited when these networks are trained with cross-entropy loss, mainly because of the unboundedness of the target function. In this paper, we aim to fill this gap by analysing the rate of the excess risk of a CNN classifier trained by cross-entropy loss. Under suitable assumptions on the smoothness and structure of the a posteriori probability, it is shown that these classifiers achieve a rate of convergence which is independent of the dimension of the image. These rates are in line with the practical observations about CNNs.
{"title":"Statistical theory for image classification using deep convolutional neural network with cross-entropy loss under the hierarchical max-pooling model","authors":"Michael Kohler , Sophie Langer","doi":"10.1016/j.jspi.2024.106188","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106188","url":null,"abstract":"<div><p>Convolutional neural networks (CNNs) trained with cross-entropy loss have proven to be extremely successful in classifying images. In recent years, much work has been done to also improve the theoretical understanding of neural networks. Nevertheless, it seems limited when these networks are trained with cross-entropy loss, mainly because of the unboundedness of the target function. In this paper, we aim to fill this gap by analysing the rate of the excess risk of a CNN classifier trained by cross-entropy loss. Under suitable assumptions on the smoothness and structure of the a posteriori probability, it is shown that these classifiers achieve a rate of convergence which is independent of the dimension of the image. These rates are in line with the practical observations about CNNs.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"234 ","pages":"Article 106188"},"PeriodicalIF":0.9,"publicationDate":"2024-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0378375824000454/pdfft?md5=68a8b5f0ef9e0563ac8f09f8ca152533&pid=1-s2.0-S0378375824000454-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141422984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-05DOI: 10.1016/j.jspi.2024.106198
Xiangyu Fang , Hongyi Li , Zujun Ou
In this paper, two simple and effective construction methods are proposed to construct four-level design with large size via quaternary codes from some small two-level initial designs. Under the popular criteria for selecting optimal design, such as generalized minimum aberration, minimum moment aberration and uniformity measured by average Lee discrepancy, the close relationships between the constructed four-level design and its initial design are investigated, which provide the guidance for choosing the suitable initial design. Moreover, some lower bounds of average Lee discrepancy for the constructed four-level designs are obtained, which can be used as a benchmark for evaluating the uniformity of the constructed four-level designs. Some numerical examples show that the large four-level designs can be constructed with high efficiency.
{"title":"Construction on large four-level designs via quaternary codes","authors":"Xiangyu Fang , Hongyi Li , Zujun Ou","doi":"10.1016/j.jspi.2024.106198","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106198","url":null,"abstract":"<div><p>In this paper, two simple and effective construction methods are proposed to construct four-level design with large size via quaternary codes from some small two-level initial designs. Under the popular criteria for selecting optimal design, such as generalized minimum aberration, minimum moment aberration and uniformity measured by average Lee discrepancy, the close relationships between the constructed four-level design and its initial design are investigated, which provide the guidance for choosing the suitable initial design. Moreover, some lower bounds of average Lee discrepancy for the constructed four-level designs are obtained, which can be used as a benchmark for evaluating the uniformity of the constructed four-level designs. Some numerical examples show that the large four-level designs can be constructed with high efficiency.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"234 ","pages":"Article 106198"},"PeriodicalIF":0.9,"publicationDate":"2024-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141302737","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-01DOI: 10.1016/j.jspi.2024.106196
Hao Zeng , Chuang Wan , Wei Zhong , Tuo Liu
Integrative analysis plays a critical role in integrating heterogeneous data from multiple datasets to provide a comprehensive view of the overall data features. However, in multiple datasets, outliers and heavy-tailed data can render least squares estimation unreliable. In response, we propose a Robust Integrative Analysis via Quantile Regression (RIAQ) that accounts for homogeneity and sparsity in multiple datasets. The RIAQ approach is not only able to identify latent homogeneous coefficient structures but also recover the sparsity of high-dimensional covariates via double penalty terms. The integration of sample information across multiple datasets improves estimation efficiency, while a sparse model improves model interpretability. Furthermore, quantile regression allows the detection of subgroup structures under different quantile levels, providing a comprehensive picture of the relationship between response and high-dimensional covariates. We develop an efficient alternating direction method of multipliers (ADMM) algorithm to solve the optimization problem and study its convergence. We also derive the parameter selection consistency of the modified Bayesian information criterion. Numerical studies demonstrate that our proposed estimator has satisfactory finite-sample performance, especially in heavy-tailed cases.
{"title":"Robust Integrative Analysis via Quantile Regression with Homogeneity and Sparsity","authors":"Hao Zeng , Chuang Wan , Wei Zhong , Tuo Liu","doi":"10.1016/j.jspi.2024.106196","DOIUrl":"10.1016/j.jspi.2024.106196","url":null,"abstract":"<div><p>Integrative analysis plays a critical role in integrating heterogeneous data from multiple datasets to provide a comprehensive view of the overall data features. However, in multiple datasets, outliers and heavy-tailed data can render least squares estimation unreliable. In response, we propose a Robust Integrative Analysis via Quantile Regression (RIAQ) that accounts for homogeneity and sparsity in multiple datasets. The RIAQ approach is not only able to identify latent homogeneous coefficient structures but also recover the sparsity of high-dimensional covariates via double penalty terms. The integration of sample information across multiple datasets improves estimation efficiency, while a sparse model improves model interpretability. Furthermore, quantile regression allows the detection of subgroup structures under different quantile levels, providing a comprehensive picture of the relationship between response and high-dimensional covariates. We develop an efficient alternating direction method of multipliers (ADMM) algorithm to solve the optimization problem and study its convergence. We also derive the parameter selection consistency of the modified Bayesian information criterion. Numerical studies demonstrate that our proposed estimator has satisfactory finite-sample performance, especially in heavy-tailed cases.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"234 ","pages":"Article 106196"},"PeriodicalIF":0.9,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141282198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-28DOI: 10.1016/j.jspi.2024.106194
Anne-Marie Toparkus, Rafael Weißbach
In studies on lifetimes, occasionally, the population contains statistical units that are born before the data collection has started. Left-truncated are units that deceased before this start. For all other units, the age at the study start often is recorded and we aim at testing whether this second measurement is independent of the genuine measure of interest, the lifetime. Our basic model of dependence is the one-parameter Gumbel–Barnett copula. For simplicity, the marginal distribution of the lifetime is assumed to be Exponential and for the age-at-study-start, namely the distribution of birth dates, we assume a Uniform. Also for simplicity, and to fit our application, we assume that units that die later than our study period, are also truncated. As a result from point process theory, we can approximate the truncated sample by a Poisson process and thereby derive its likelihood. Identification, consistency and asymptotic distribution of the maximum-likelihood estimator are derived. Testing for positive truncation dependence must include the hypothetical independence which coincides with the boundary of the copula’s parameter space. By non-standard theory, the maximum likelihood estimator of the exponential and the copula parameter is distributed as a mixture of a two- and a one-dimensional normal distribution. For the proof, the third parameter, the unobservable sample size, is profiled out. An interesting result is, that it differs to view the data as truncated sample, or, as simple sample from the truncated population, but not by much. The application are 55 thousand double-truncated lifetimes of German businesses that closed down over the period 2014 to 2016. The likelihood has its maximum for the copula parameter at the parameter space boundary so that the -value of test is 0.5. The life expectancy does not increase relative to the year of foundation. Using a Farlie–Gumbel–Morgenstern copula, which models positive and negative dependence, finds that life expectancy of German enterprises even decreases significantly over time. A simulation under the condition of the application suggests that the tests retain the nominal level and have good power.
{"title":"Testing truncation dependence: The Gumbel–Barnett copula","authors":"Anne-Marie Toparkus, Rafael Weißbach","doi":"10.1016/j.jspi.2024.106194","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106194","url":null,"abstract":"<div><p>In studies on lifetimes, occasionally, the population contains statistical units that are born before the data collection has started. Left-truncated are units that deceased before this start. For all other units, the age at the study start often is recorded and we aim at testing whether this second measurement is independent of the genuine measure of interest, the lifetime. Our basic model of dependence is the one-parameter Gumbel–Barnett copula. For simplicity, the marginal distribution of the lifetime is assumed to be Exponential and for the age-at-study-start, namely the distribution of birth dates, we assume a Uniform. Also for simplicity, and to fit our application, we assume that units that die later than our study period, are also truncated. As a result from point process theory, we can approximate the truncated sample by a Poisson process and thereby derive its likelihood. Identification, consistency and asymptotic distribution of the maximum-likelihood estimator are derived. Testing for positive truncation dependence must include the hypothetical independence which coincides with the boundary of the copula’s parameter space. By non-standard theory, the maximum likelihood estimator of the exponential and the copula parameter is distributed as a mixture of a two- and a one-dimensional normal distribution. For the proof, the third parameter, the unobservable sample size, is profiled out. An interesting result is, that it differs to view the data as truncated sample, or, as simple sample from the truncated population, but not by much. The application are 55 thousand double-truncated lifetimes of German businesses that closed down over the period 2014 to 2016. The likelihood has its maximum for the copula parameter at the parameter space boundary so that the <span><math><mi>p</mi></math></span>-value of test is 0.5. The life expectancy does not increase relative to the year of foundation. Using a Farlie–Gumbel–Morgenstern copula, which models positive and negative dependence, finds that life expectancy of German enterprises even decreases significantly over time. A simulation under the condition of the application suggests that the tests retain the nominal level and have good power.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"234 ","pages":"Article 106194"},"PeriodicalIF":0.9,"publicationDate":"2024-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S037837582400051X/pdfft?md5=a5bc737bb68bd11a1a31f4aeb333c40e&pid=1-s2.0-S037837582400051X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141240222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-21DOI: 10.1016/j.jspi.2024.106192
Yingnan Zhang, Jiangmin Pan, Lei Shi
Row–column designs that provide unconfounded estimation of all main effects and the maximum number of two-factor interactions (2fi’s) are called 2fi-optimal. This issue has been paid great attention recently for its wide application in industrial or physical experiments. The constructions of 2fi-optimal two-level and three-level full factorial and fractional factorial row–column designs have been proposed. However, the results for higher prime levels have not been achieved yet. In this paper, we give theoretical constructions of 2fi-optimal full factorial row–column designs for any odd prime level and any parameter combination, and theoretical constructions of 2fi-optimal fractional factorial row–column designs for any prime level and any parameter combination.
能对所有主效应和最大数量的双因素交互作用(2fi)进行无约束估计的行列式设计被称为 2fi 最佳设计。最近,这一问题因其在工业或物理实验中的广泛应用而备受关注。已有人提出了 2fi-optimal 两级和三级全因子和分数因子行列式设计的构造。但是,对于更高的素数级,目前还没有结果。在本文中,我们给出了针对任意奇数素数级 s 和任意参数组合的 2fi-optimal sn 全因子行列式设计的理论构造,以及针对任意素数级 s 和任意参数组合的 2fi-optimal sn-1 小数因子行列式设计的理论构造。
{"title":"Construction of 2fi-optimal row–column designs","authors":"Yingnan Zhang, Jiangmin Pan, Lei Shi","doi":"10.1016/j.jspi.2024.106192","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106192","url":null,"abstract":"<div><p>Row–column designs that provide unconfounded estimation of all main effects and the maximum number of two-factor interactions (2fi’s) are called 2fi-optimal. This issue has been paid great attention recently for its wide application in industrial or physical experiments. The constructions of 2fi-optimal two-level and three-level full factorial and fractional factorial row–column designs have been proposed. However, the results for higher prime levels have not been achieved yet. In this paper, we give theoretical constructions of 2fi-optimal <span><math><msup><mrow><mi>s</mi></mrow><mrow><mi>n</mi></mrow></msup></math></span> full factorial row–column designs for any odd prime level <span><math><mi>s</mi></math></span> and any parameter combination, and theoretical constructions of 2fi-optimal <span><math><msup><mrow><mi>s</mi></mrow><mrow><mi>n</mi><mo>−</mo><mn>1</mn></mrow></msup></math></span> fractional factorial row–column designs for any prime level <span><math><mi>s</mi></math></span> and any parameter combination.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"234 ","pages":"Article 106192"},"PeriodicalIF":0.9,"publicationDate":"2024-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141164387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-21DOI: 10.1016/j.jspi.2024.106193
Francisco Cribari-Neto, José Jairo Santana-e-Silva, Klaus L.P. Vasconcellos
The beta regression model is tailored for responses that assume values in the standard unit interval. It comprises two submodels, one for the mean response and another for the precision parameter. We develop tests of correct specification for such a model. The tests are based on the information matrix equality, which holds when the model is correctly specified. We establish the validity of the tests in the class of varying precision beta regressions, provide closed-form expressions for the quantities used in the test statistics, and present simulation evidence on the tests’ null and non-null behavior. We show that it is possible to achieve very good control of the type I error probability when data resampling is employed and that the tests are able to reliably detect incorrect model specification, especially when the sample size is not small. An empirical application is presented and discussed.
贝塔回归模型适用于在标准单位区间内取值的响应。它包括两个子模型,一个用于平均响应,另一个用于精确参数。我们对这种模型的正确规范进行了检验。这些检验以信息矩阵相等为基础,当模型被正确指定时,信息矩阵相等成立。我们确定了检验在不同精度贝塔回归类中的有效性,提供了检验统计中使用的量的闭式表达式,并提出了检验的无效和非无效行为的模拟证据。我们表明,在采用数据重采样时,可以很好地控制 I 类错误概率,而且检验能够可靠地检测出错误的模型规范,尤其是在样本量不小的情况下。本文介绍并讨论了一个经验应用。
{"title":"Beta regression misspecification tests","authors":"Francisco Cribari-Neto, José Jairo Santana-e-Silva, Klaus L.P. Vasconcellos","doi":"10.1016/j.jspi.2024.106193","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106193","url":null,"abstract":"<div><p>The beta regression model is tailored for responses that assume values in the standard unit interval. It comprises two submodels, one for the mean response and another for the precision parameter. We develop tests of correct specification for such a model. The tests are based on the information matrix equality, which holds when the model is correctly specified. We establish the validity of the tests in the class of varying precision beta regressions, provide closed-form expressions for the quantities used in the test statistics, and present simulation evidence on the tests’ null and non-null behavior. We show that it is possible to achieve very good control of the type I error probability when data resampling is employed and that the tests are able to reliably detect incorrect model specification, especially when the sample size is not small. An empirical application is presented and discussed.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"233 ","pages":"Article 106193"},"PeriodicalIF":0.9,"publicationDate":"2024-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141096512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-15DOI: 10.1016/j.jspi.2024.106191
Richeng Hu , Ngai-Hang Chan , Rongmao Zhang
A self-normalized approach for testing the stationarity of a -dimensional random field is considered in this paper. Because the discrete Fourier transforms (DFT) at fundamental frequencies of a second-order stationary random field are asymptotically uncorrelated (see Bandyopadhyay and Subba Rao, 2017), one can construct a stationarity test based on the sample covariance of the DFTs. Such a test is usually inferior because it involves an overestimated scale parameter that leads to low size and power. To circumvent this shortcoming, this paper proposes two self-normalized statistics based on extreme value and partial sum of the sample covariance of the DFTs. Under certain regularity conditions, it is shown that the proposed tests converge to functionals of Brownian motion. Simulations and a data analysis demonstrate the outstanding performance of the proposed tests.
{"title":"Self-normalized inference for stationarity of irregular spatial data","authors":"Richeng Hu , Ngai-Hang Chan , Rongmao Zhang","doi":"10.1016/j.jspi.2024.106191","DOIUrl":"10.1016/j.jspi.2024.106191","url":null,"abstract":"<div><p>A self-normalized approach for testing the stationarity of a <span><math><mi>d</mi></math></span>-dimensional random field is considered in this paper. Because the discrete Fourier transforms (DFT) at fundamental frequencies of a second-order stationary random field are asymptotically uncorrelated (see Bandyopadhyay and Subba Rao, 2017), one can construct a stationarity test based on the sample covariance of the DFTs. Such a test is usually inferior because it involves an overestimated scale parameter that leads to low size and power. To circumvent this shortcoming, this paper proposes two self-normalized statistics based on extreme value and partial sum of the sample covariance of the DFTs. Under certain regularity conditions, it is shown that the proposed tests converge to functionals of Brownian motion. Simulations and a data analysis demonstrate the outstanding performance of the proposed tests.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"234 ","pages":"Article 106191"},"PeriodicalIF":0.9,"publicationDate":"2024-05-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141046356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-10DOI: 10.1016/j.jspi.2024.106189
Michaël Allouche , Jonathan El Methni , Stéphane Girard
Conditional tail expectation (CTE) is a coherent risk measure defined as the mean of the loss distribution above a high quantile. The existence of the CTE as well as the asymptotic properties of associated estimators however require integrability conditions that may be violated when dealing with heavy-tailed distributions. We introduce Box–Cox transforms of the CTE that have two benefits. First, they alleviate these theoretical issues. Second, they enable to recover a number of risk measures such as conditional tail expectation, expected shortfall, conditional value-at-risk or conditional tail variance. The construction of dedicated estimators is based on the investigation of the asymptotic relationship between Box–Cox transforms of the CTE and quantiles at extreme probability levels, as well as on an extrapolation formula established in the heavy-tailed context. We quantify and estimate the bias induced by the use of these approximations and then introduce reduced-bias estimators whose asymptotic properties are rigorously shown. Their finite-sample properties are assessed on a simulation study and illustrated on real data, highlighting the practical interest of both the bias reduction and the Box–Cox transform.
{"title":"Reduced-bias estimation of the extreme conditional tail expectation for Box–Cox transforms of heavy-tailed distributions","authors":"Michaël Allouche , Jonathan El Methni , Stéphane Girard","doi":"10.1016/j.jspi.2024.106189","DOIUrl":"10.1016/j.jspi.2024.106189","url":null,"abstract":"<div><p>Conditional tail expectation (CTE) is a coherent risk measure defined as the mean of the loss distribution above a high quantile. The existence of the CTE as well as the asymptotic properties of associated estimators however require integrability conditions that may be violated when dealing with heavy-tailed distributions. We introduce Box–Cox transforms of the CTE that have two benefits. First, they alleviate these theoretical issues. Second, they enable to recover a number of risk measures such as conditional tail expectation, expected shortfall, conditional value-at-risk or conditional tail variance. The construction of dedicated estimators is based on the investigation of the asymptotic relationship between Box–Cox transforms of the CTE and quantiles at extreme probability levels, as well as on an extrapolation formula established in the heavy-tailed context. We quantify and estimate the bias induced by the use of these approximations and then introduce reduced-bias estimators whose asymptotic properties are rigorously shown. Their finite-sample properties are assessed on a simulation study and illustrated on real data, highlighting the practical interest of both the bias reduction and the Box–Cox transform.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"233 ","pages":"Article 106189"},"PeriodicalIF":0.9,"publicationDate":"2024-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141035013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}