Pub Date : 2023-11-18DOI: 10.1016/j.jspi.2023.106122
Isadora Antoniano-Villalobos , Cristiano Villa , Stephen G. Walker
The construction of objective priors is, at best, challenging for multidimensional parameter spaces. A common practice is to assume independence and set up the joint prior as the product of marginal distributions obtained via “standard” objective methods, such as Jeffreys or reference priors. However, the assumption of independence a priori is not always reasonable, and whether it can be viewed as strictly objective is still open to discussion. In this paper, by extending a previously proposed objective approach based on scoring rules for the one dimensional case, we propose a novel objective prior for multidimensional parameter spaces which yields a dependence structure. The proposed prior has the appealing property of being proper and does not depend on the chosen model; only on the parameter space considered.
{"title":"A multidimensional objective prior distribution from a scoring rule","authors":"Isadora Antoniano-Villalobos , Cristiano Villa , Stephen G. Walker","doi":"10.1016/j.jspi.2023.106122","DOIUrl":"https://doi.org/10.1016/j.jspi.2023.106122","url":null,"abstract":"<div><p>The construction of objective priors is, at best, challenging for multidimensional parameter spaces. A common practice is to assume independence and set up the joint prior as the product of marginal distributions obtained via “standard” objective methods, such as Jeffreys or reference priors. However, the assumption of independence a priori is not always reasonable, and whether it can be viewed as strictly objective is still open to discussion. In this paper, by extending a previously proposed objective approach based on scoring rules for the one dimensional case, we propose a novel objective prior for multidimensional parameter spaces which yields a dependence structure. The proposed prior has the appealing property of being proper and does not depend on the chosen model; only on the parameter space considered.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"231 ","pages":"Article 106122"},"PeriodicalIF":0.9,"publicationDate":"2023-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138395335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-11-15DOI: 10.1016/j.jspi.2023.106120
Yifan Sun , Ziyi Liu , Wu Wang
Classical functional linear regression models the relationship between a scalar response and a functional covariate, where the coefficient function is assumed to be identical for all subjects. In this paper, the classical model is extended to allow heterogeneous coefficient functions across different subgroups of subjects. The greatest challenge is that the subgroup structure is usually unknown to us. To this end, we develop a penalization-based approach which innovatively applies the penalized fusion technique to simultaneously determine the number and structure of subgroups and coefficient functions within each subgroup. An effective computational algorithm is derived. We also establish the oracle properties and estimation consistency. Extensive numerical simulations demonstrate its superiority compared to several competing methods. The analysis of an air quality dataset leads to interesting findings and improved predictions.
{"title":"Subgroup analysis for the functional linear model","authors":"Yifan Sun , Ziyi Liu , Wu Wang","doi":"10.1016/j.jspi.2023.106120","DOIUrl":"https://doi.org/10.1016/j.jspi.2023.106120","url":null,"abstract":"<div><p>Classical functional linear regression models the relationship between a scalar response and a functional covariate, where the coefficient function is assumed to be identical for all subjects. In this paper, the classical model is extended to allow heterogeneous coefficient functions across different subgroups of subjects. The greatest challenge is that the subgroup structure is usually unknown to us. To this end, we develop a penalization-based approach which innovatively applies the penalized fusion technique to simultaneously determine the number and structure of subgroups and coefficient functions within each subgroup. An effective computational algorithm is derived. We also establish the oracle properties and estimation consistency. Extensive numerical simulations demonstrate its superiority compared to several competing methods. The analysis of an air quality dataset leads to interesting findings and improved predictions.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"231 ","pages":"Article 106120"},"PeriodicalIF":0.9,"publicationDate":"2023-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138435832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-11-13DOI: 10.1016/j.jspi.2023.106119
Mehrdad Pournaderi, Yu Xiang
The fixed-X knockoff filter is a flexible framework for variable selection with false discovery rate (FDR) control in linear models with arbitrary design matrices (of full column rank) and it allows for finite-sample selective inference via the Lasso estimates. In this paper, we extend the theory of the knockoff procedure to tests with composite null hypotheses, which are usually more relevant to real-world problems. The main technical challenge lies in handling composite nulls in tandem with dependent features from arbitrary designs. We develop two methods for composite inference with the knockoffs, namely, shifted ordinary least-squares (S-OLS) and feature-response product perturbation (FRPP), building on new structural properties of test statistics under composite nulls. We also propose two heuristic variants of S-OLS method that outperform the celebrated Benjamini–Hochberg (BH) procedure for composite nulls, which serves as a heuristic baseline under dependent test statistics. Finally, we analyze the loss in FDR when the original knockoff procedure is naively applied on composite tests.
{"title":"Variable selection with the knockoffs: Composite null hypotheses","authors":"Mehrdad Pournaderi, Yu Xiang","doi":"10.1016/j.jspi.2023.106119","DOIUrl":"10.1016/j.jspi.2023.106119","url":null,"abstract":"<div><p>The fixed-X knockoff filter is a flexible framework for variable selection with false discovery rate<span> (FDR) control in linear models with arbitrary design matrices<span><span> (of full column rank) and it allows for finite-sample selective inference via the Lasso estimates. In this paper, we extend the theory of the knockoff procedure to tests with composite null hypotheses, which are usually more relevant to real-world problems. The main technical challenge lies in handling composite </span>nulls in tandem with dependent features from arbitrary designs. We develop two methods for composite inference with the knockoffs, namely, shifted ordinary least-squares (S-OLS) and feature-response product perturbation (FRPP), building on new structural properties of test statistics under composite nulls. We also propose two heuristic variants of S-OLS method that outperform the celebrated Benjamini–Hochberg (BH) procedure for composite nulls, which serves as a heuristic baseline under dependent test statistics. Finally, we analyze the loss in FDR when the original knockoff procedure is naively applied on composite tests.</span></span></p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"231 ","pages":"Article 106119"},"PeriodicalIF":0.9,"publicationDate":"2023-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135714688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-11-10DOI: 10.1016/j.jspi.2023.106118
Hui Li , Liuqing Yang , Kashinath Chatterjee , Min-Qian Liu
Supersaturated design (SSD) plays an important role in screening factors. criterion is one of the most widely used criteria to evaluate multi-level and mixed-level SSDs. This paper provides some methods to construct multi-level optimal SSDs with general run sizes, which can also be extended to construct mixed-level SSDs. The main idea of these methods is combining two processed generalized Hadamard matrices with the expansive replacement method. These proposed methods are easy to perform. Besides, the non-orthogonality between any two columns of the resulting SSDs is well controlled by that of the source designs. Some illustrative examples are given and several new SSDs are provided in this paper.
{"title":"Construction of optimal supersaturated designs by the expansive replacement method","authors":"Hui Li , Liuqing Yang , Kashinath Chatterjee , Min-Qian Liu","doi":"10.1016/j.jspi.2023.106118","DOIUrl":"https://doi.org/10.1016/j.jspi.2023.106118","url":null,"abstract":"<div><p>Supersaturated design (SSD) plays an important role in screening factors. <span><math><mrow><mi>E</mi><mrow><mo>(</mo><msub><mrow><mi>f</mi></mrow><mrow><mi>N</mi><mi>O</mi><mi>D</mi></mrow></msub><mo>)</mo></mrow></mrow></math></span> criterion is one of the most widely used criteria to evaluate multi-level and mixed-level SSDs. This paper provides some methods to construct multi-level <span><math><mrow><mi>E</mi><mrow><mo>(</mo><msub><mrow><mi>f</mi></mrow><mrow><mi>N</mi><mi>O</mi><mi>D</mi></mrow></msub><mo>)</mo></mrow></mrow></math></span> optimal SSDs with general run sizes, which can also be extended to construct mixed-level SSDs. The main idea of these methods is combining two processed generalized Hadamard matrices with the expansive replacement method. These proposed methods are easy to perform. Besides, the non-orthogonality between any two columns of the resulting SSDs is well controlled by that of the source designs. Some illustrative examples are given and several new SSDs are provided in this paper.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"230 ","pages":"Article 106118"},"PeriodicalIF":0.9,"publicationDate":"2023-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"92043598","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-11-10DOI: 10.1016/j.jspi.2023.106117
Takemi Yanagimoto , Yoichi Miyata
A Bayesian estimator aiming at improving the conditional MLE is proposed by introducing a pair of priors. After explaining the conditional MLE by the posterior mode under a prior, we define a promising estimator by the posterior mean under a corresponding prior. The prior is asymptotically equivalent to the reference prior in familiar models. Advantages of the present approach include two different optimality properties of the induced estimator, the ease of various extensions and the possible treatments for a finite sample size. The existing approaches are discussed and critiqued.
{"title":"A pair of novel priors for improving and extending the conditional MLE","authors":"Takemi Yanagimoto , Yoichi Miyata","doi":"10.1016/j.jspi.2023.106117","DOIUrl":"10.1016/j.jspi.2023.106117","url":null,"abstract":"<div><p><span>A Bayesian estimator aiming at improving the conditional MLE is proposed by introducing a pair of priors. After explaining the conditional MLE by the posterior mode under a prior, we define a promising estimator by the </span>posterior mean<span> under a corresponding prior. The prior is asymptotically equivalent to the reference prior in familiar models. Advantages of the present approach include two different optimality properties of the induced estimator, the ease of various extensions and the possible treatments for a finite sample size. The existing approaches are discussed and critiqued.</span></p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"231 ","pages":"Article 106117"},"PeriodicalIF":0.9,"publicationDate":"2023-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135615023","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-10-26DOI: 10.1016/j.jspi.2023.106116
Rosanne J. Turner , Alexander Ly , Peter D. Grünwald
We develop -variables for testing whether two or more data streams come from the same source or not, and more generally, whether the difference between the sources is larger than some minimal effect size. These -variables lead to exact, nonasymptotic tests that remain safe, i.e., keep their type-I error guarantees, under flexible sampling scenarios such as optional stopping and continuation. In special cases our -variables also have an optimal ‘growth’ property under the alternative. While the construction is generic, we illustrate it through the special case of contingency tables, i.e. Bernoulli streams, allowing for the incorporation of different restrictions on the composite alternative. Comparison to -value analysis in simulations and a real-world 2 × 2 contingency table example show that -variables, through their flexibility, often allow for early stopping of data collection — thereby retaining similar power as classical methods — while also retaining the option of extending or combining data afterwards.
{"title":"Generic E-variables for exact sequential k-sample tests that allow for optional stopping","authors":"Rosanne J. Turner , Alexander Ly , Peter D. Grünwald","doi":"10.1016/j.jspi.2023.106116","DOIUrl":"https://doi.org/10.1016/j.jspi.2023.106116","url":null,"abstract":"<div><p>We develop <span><math><mstyle><mi>E</mi></mstyle></math></span>-variables for testing whether two or more data streams come from the same source or not, and more generally, whether the difference between the sources is larger than some minimal effect size. These <span><math><mstyle><mi>E</mi></mstyle></math></span>-variables lead to exact, nonasymptotic tests that remain safe, i.e., keep their type-I error guarantees, under flexible sampling scenarios such as optional stopping and continuation. In special cases our <span><math><mstyle><mi>E</mi></mstyle></math></span>-variables also have an optimal ‘growth’ property under the alternative. While the construction is generic, we illustrate it through the special case of <span><math><mrow><mi>k</mi><mo>×</mo><mn>2</mn></mrow></math></span> contingency tables, i.e. <span><math><mi>k</mi></math></span> Bernoulli streams, allowing for the incorporation of different restrictions on the composite alternative. Comparison to <span><math><mi>p</mi></math></span>-value analysis in simulations and a real-world 2 × 2 contingency table example show that <span><math><mstyle><mi>E</mi></mstyle></math></span>-variables, through their flexibility, often allow for early stopping of data collection — thereby retaining similar power as classical methods — while also retaining the option of extending or combining data afterwards.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"230 ","pages":"Article 106116"},"PeriodicalIF":0.9,"publicationDate":"2023-10-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S037837582300085X/pdfft?md5=572bc8e92c25baa3e6a3f4936ee83e72&pid=1-s2.0-S037837582300085X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"92043596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-10-13DOI: 10.1016/j.jspi.2023.106115
Victoria L. Leaver , Robert G. Clark , Pavel N. Krivitsky , Carole L. Birrell
Three likelihood approaches to estimation under informative sampling are compared using a special case for which analytic expressions are possible to derive. An independent and identically distributed population of values of a variable of interest is drawn from a gamma distribution, with the shape parameter and the population size both assumed to be known. The sampling method is selection with probability proportional to a power of the variable with replacement, so that duplicate sample units are possible. Estimators of the unknown parameter, variance estimators and asymptotic variances of the estimators are derived for maximum likelihood, sample likelihood and pseudo-likelihood estimation. Theoretical derivations and simulation results show that the efficiency of the sample likelihood approaches that of full maximum likelihood estimation when the sample size tends to infinity and the sampling fraction tends to zero. However, when tends to infinity and is not negligible, the maximum likelihood estimator is more efficient than the other methods because it takes the possibility of duplicate sample units into account. Pseudo-likelihood can perform much more poorly than the other methods in some cases. For the special case when the superpopulation is exponential and the selection is probability proportional to size, the anticipated variance of the pseudo-likelihood estimate is infinite.
{"title":"A comparison of likelihood-based methods for size-biased sampling","authors":"Victoria L. Leaver , Robert G. Clark , Pavel N. Krivitsky , Carole L. Birrell","doi":"10.1016/j.jspi.2023.106115","DOIUrl":"https://doi.org/10.1016/j.jspi.2023.106115","url":null,"abstract":"<div><p>Three likelihood approaches to estimation under informative sampling are compared using a special case for which analytic expressions are possible to derive. An independent and identically distributed population of values of a variable of interest is drawn from a gamma distribution, with the shape parameter and the population size both assumed to be known. The sampling method is selection with probability proportional to a power of the variable with replacement, so that duplicate sample units are possible. Estimators of the unknown parameter, variance estimators and asymptotic variances of the estimators are derived for maximum likelihood, sample likelihood and pseudo-likelihood estimation. Theoretical derivations and simulation results show that the efficiency of the sample likelihood approaches that of full maximum likelihood estimation when the sample size <span><math><mi>n</mi></math></span> tends to infinity and the sampling fraction <span><math><mi>f</mi></math></span> tends to zero. However, when <span><math><mi>n</mi></math></span> tends to infinity and <span><math><mi>f</mi></math></span> is not negligible, the maximum likelihood estimator is more efficient than the other methods because it takes the possibility of duplicate sample units into account. Pseudo-likelihood can perform much more poorly than the other methods in some cases. For the special case when the superpopulation is exponential and the selection is probability proportional to size, the anticipated variance of the pseudo-likelihood estimate is infinite.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"230 ","pages":"Article 106115"},"PeriodicalIF":0.9,"publicationDate":"2023-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0378375823000848/pdfft?md5=34807a0d3caadad51aaee1e2b82b751e&pid=1-s2.0-S0378375823000848-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"92115764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-10-07DOI: 10.1016/j.jspi.2023.106113
Fadoua Balabdaoui, Harald Besdziek
The two-component mixture model with known background density, unknown signal density, and unknown mixing proportion has been studied in many contexts. One such context is multiple testing, where the background and signal densities describe the distribution of the -values under the null and alternative hypotheses, respectively. In this paper, we consider the log-concave MLE of the signal density using the estimator of Patra & Sen (2016) for the mixing probability. We show that it is consistent and converges at the global rate . An EM-algorithm in combination with an active set algorithm implemented in the R-package logcondens was used to compute the log-concave MLE. When one is interested in estimation at a fixed point, a conjecture is made about the limit distribution of our estimator. The performance of our method is assessed through a simulation study.
{"title":"Maximum likelihood estimation of the log-concave component in a semi-parametric mixture with a standard normal density","authors":"Fadoua Balabdaoui, Harald Besdziek","doi":"10.1016/j.jspi.2023.106113","DOIUrl":"https://doi.org/10.1016/j.jspi.2023.106113","url":null,"abstract":"<div><p>The two-component mixture model with known background density, unknown signal density, and unknown mixing proportion has been studied in many contexts. One such context is multiple testing, where the background and signal densities describe the distribution of the <span><math><mi>p</mi></math></span><span>-values under the null and alternative hypotheses, respectively. In this paper, we consider the log-concave MLE of the signal density using the estimator of Patra & Sen (2016) for the mixing probability. We show that it is consistent and converges at the global rate </span><span><math><msup><mrow><mi>n</mi></mrow><mrow><mo>−</mo><mn>2</mn><mo>/</mo><mn>5</mn></mrow></msup></math></span>. An EM-algorithm in combination with an active set algorithm implemented in the R-package logcondens was used to compute the log-concave MLE. When one is interested in estimation at a fixed point, a conjecture is made about the limit distribution of our estimator. The performance of our method is assessed through a simulation study.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"230 ","pages":"Article 106113"},"PeriodicalIF":0.9,"publicationDate":"2023-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49903459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-09-27DOI: 10.1016/j.jspi.2023.106114
Juan José Fernández-Durán, María Mercedes Gregorio-Domínguez
The parameter space of nonnegative trigonometric sums (NNTS) models for circular data is the surface of a hypersphere; thus, constructing regression models for a circular-dependent variable using NNTS models can comprise fitting great (small) circles on the parameter hypersphere that can identify different regions (rotations) along the great (small) circle. We propose regression models for circular- (angular-) dependent random variables in which the original circular random variable, which is assumed to be distributed (marginally) as an NNTS model, is transformed into a linear random variable such that common methods for linear regression can be applied. The usefulness of NNTS models with skewness and multimodality is shown in examples with simulated and real data.
{"title":"Regression models for circular data based on nonnegative trigonometric sums","authors":"Juan José Fernández-Durán, María Mercedes Gregorio-Domínguez","doi":"10.1016/j.jspi.2023.106114","DOIUrl":"https://doi.org/10.1016/j.jspi.2023.106114","url":null,"abstract":"<div><p>The parameter space of nonnegative trigonometric sums (NNTS) models for circular data is the surface of a hypersphere<span>; thus, constructing regression models for a circular-dependent variable using NNTS models can comprise fitting great (small) circles on the parameter hypersphere that can identify different regions (rotations) along the great (small) circle. We propose regression models for circular- (angular-) dependent random variables in which the original circular random variable, which is assumed to be distributed (marginally) as an NNTS model, is transformed into a linear random variable such that common methods for linear regression can be applied. The usefulness of NNTS models with skewness and multimodality is shown in examples with simulated and real data.</span></p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"230 ","pages":"Article 106114"},"PeriodicalIF":0.9,"publicationDate":"2023-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49903457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-09-23DOI: 10.1016/j.jspi.2023.106112
V. Girardin , R. Senoussi
A scalar discrete or continuous time process is reducible to stationarity (RWS) if its transform by some smooth time change is weakly stationary. Different issues linked to this notion are here investigated for autoregressive (AR) models. AR models are understood in a large sense and may have time-varying coefficients. In the continuous time case the innovation may be of the semi-martingale type–such as compound Poisson noise; in the discrete case, the noise may not be Gaussian.
Necessary and sufficient conditions for scalar AR models to be RWS are investigated, with explicit formulas for the time changes. Stationarity reduction issues for discrete sequences sampled from time continuous AR processes are also considered. Several types of time changes, RWS processes and sequences are studied with examples and simulation, including the classical multiplicative stationary AR models.
{"title":"Time changes and stationarity issues for extended scalar autoregressive models","authors":"V. Girardin , R. Senoussi","doi":"10.1016/j.jspi.2023.106112","DOIUrl":"https://doi.org/10.1016/j.jspi.2023.106112","url":null,"abstract":"<div><p>A scalar discrete or continuous time process is reducible to stationarity (RWS) if its transform by some smooth time change is weakly stationary. Different issues linked to this notion are here investigated for autoregressive (AR) models. AR models are understood in a large sense and may have time-varying coefficients. In the continuous time case the innovation may be of the semi-martingale type–such as compound Poisson noise; in the discrete case, the noise may not be Gaussian.</p><p>Necessary and sufficient conditions for scalar AR models to be RWS are investigated, with explicit formulas for the time changes. Stationarity reduction issues for discrete sequences sampled from time continuous AR processes are also considered. Several types of time changes, RWS processes and sequences are studied with examples and simulation, including the classical multiplicative stationary AR models.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"230 ","pages":"Article 106112"},"PeriodicalIF":0.9,"publicationDate":"2023-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49903458","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}