Pub Date : 2024-02-28DOI: 10.1016/j.jspi.2024.106164
Wanfeng Liang , Yuhao Zhang , Jiyang Wang , Yue Wu , Xiaoyan Ma
The modified Cholesky decomposition (MCD) method is commonly used in precision matrix estimation assuming that the random variables have a specified order. In this paper, we develop a permutation-based refitted cross validation (PRCV) estimation procedure for ultrahigh dimensional precision matrix based on the MCD, which does not rely on the order of variables. The consistency of the proposed estimator is established under the Frobenius norm without normal distribution assumption. Simulation studies present satisfactory performance of in various scenarios. The proposed method is also applied to analyze a real data. We provide the complete code at https://github.com/lwfwhunanhero/PRCV.
{"title":"A new approach for ultrahigh dimensional precision matrix estimation","authors":"Wanfeng Liang , Yuhao Zhang , Jiyang Wang , Yue Wu , Xiaoyan Ma","doi":"10.1016/j.jspi.2024.106164","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106164","url":null,"abstract":"<div><p>The modified Cholesky decomposition (MCD) method is commonly used in precision matrix estimation assuming that the random variables have a specified order. In this paper, we develop a permutation-based refitted cross validation (PRCV) estimation procedure for ultrahigh dimensional precision matrix based on the MCD, which does not rely on the order of variables. The consistency of the proposed estimator is established under the Frobenius norm without normal distribution assumption. Simulation studies present satisfactory performance of in various scenarios. The proposed method is also applied to analyze a real data. We provide the complete code at <span>https://github.com/lwfwhunanhero/PRCV</span><svg><path></path></svg>.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106164"},"PeriodicalIF":0.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139999247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-28DOI: 10.1016/j.jspi.2024.106163
William Kengne, Modou Wade
In this paper, we perform deep neural networks for learning stationary -weakly dependent processes. Such weak-dependence property includes a class of weak dependence conditions such as mixing, association and the setting considered here covers many commonly used situations such as: regression estimation, time series prediction, time series classification The consistency of the empirical risk minimization algorithm in the class of deep neural networks predictors is established. We achieve the generalization bound and obtain an asymptotic learning rate, which is less than , for all . A bound of the excess risk, for a wide class of target functions, is also derived. Applications to binary time series classification and prediction in affine causal models with exogenous covariates are carried out. Some simulation results are provided, as well as an application to the US recession data.
{"title":"Deep learning for ψ-weakly dependent processes","authors":"William Kengne, Modou Wade","doi":"10.1016/j.jspi.2024.106163","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106163","url":null,"abstract":"<div><p>In this paper, we perform deep neural networks for learning stationary <span><math><mi>ψ</mi></math></span>-weakly dependent processes. Such weak-dependence property includes a class of weak dependence conditions such as mixing, association<span><math><mrow><mo>⋯</mo><mspace></mspace></mrow></math></span> and the setting considered here covers many commonly used situations such as: regression estimation, time series prediction, time series classification<span><math><mrow><mo>⋯</mo><mspace></mspace></mrow></math></span> The consistency of the empirical risk minimization algorithm in the class of deep neural networks predictors is established. We achieve the generalization bound and obtain an asymptotic learning rate, which is less than <span><math><mrow><mi>O</mi><mrow><mo>(</mo><msup><mrow><mi>n</mi></mrow><mrow><mo>−</mo><mn>1</mn><mo>/</mo><mi>α</mi></mrow></msup><mo>)</mo></mrow></mrow></math></span>, for all <span><math><mrow><mi>α</mi><mo>></mo><mn>2</mn></mrow></math></span>. A bound of the excess risk, for a wide class of target functions, is also derived. Applications to binary time series classification and prediction in affine causal models with exogenous covariates are carried out. Some simulation results are provided, as well as an application to the US recession data.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106163"},"PeriodicalIF":0.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139999248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-27DOI: 10.1016/j.jspi.2024.106162
Xiaoyan Ma , Lili Wei , Wanfeng Liang
In this paper, we study high-dimensional reduced rank regression and propose a doubly robust procedure, called , meaning concurrent robustness to both outliers in predictors and heavy-tailed random noise. The proposed method uses the composite gradient descent based algorithm to solve the nonconvex optimization problem resulting from combining Tukey’s biweight loss with spectral regularization. Both theoretical and numerical properties of are investigated. We establish non-asymptotic estimation error bounds under both the Frobenius norm and the nuclear norm in the high-dimensional setting. Simulation studies and real example show that the performance of is better than that of several existing estimation methods.
{"title":"D4R: Doubly robust reduced rank regression in high dimension","authors":"Xiaoyan Ma , Lili Wei , Wanfeng Liang","doi":"10.1016/j.jspi.2024.106162","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106162","url":null,"abstract":"<div><p>In this paper, we study high-dimensional reduced rank regression and propose a doubly robust procedure, called <span><math><mi>D4R</mi></math></span>, meaning concurrent robustness to both outliers in predictors and heavy-tailed random noise. The proposed method uses the composite gradient descent based algorithm to solve the nonconvex optimization problem resulting from combining Tukey’s biweight loss with spectral regularization. Both theoretical and numerical properties of <span><math><mi>D4R</mi></math></span> are investigated. We establish non-asymptotic estimation error bounds under both the Frobenius norm and the nuclear norm in the high-dimensional setting. Simulation studies and real example show that the performance of <span><math><mi>D4R</mi></math></span> is better than that of several existing estimation methods.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106162"},"PeriodicalIF":0.9,"publicationDate":"2024-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139985505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-16DOI: 10.1016/j.jspi.2024.106160
Markus Kuba , Alois Panholzer
We consider a card guessing strategy for a stack of cards with two different types of cards, say cards of type red (heart or diamond) and cards of type black (clubs or spades). Given a deck of cards, we propose a refined counting of the number of correct colour guesses, when the guesser is provided with complete information, in other words, when the numbers and and the colour of each drawn card are known. We decompose the correct guessed cards into three different types by taking into account the probability of making a correct guess, and provide joint distributional results for the underlying random variables as well as joint limit laws.
{"title":"On card guessing with two types of cards","authors":"Markus Kuba , Alois Panholzer","doi":"10.1016/j.jspi.2024.106160","DOIUrl":"10.1016/j.jspi.2024.106160","url":null,"abstract":"<div><p>We consider a card guessing strategy for a stack of cards with two different types of cards, say <span><math><msub><mrow><mi>m</mi></mrow><mrow><mn>1</mn></mrow></msub></math></span> cards of type red (heart or diamond) and <span><math><msub><mrow><mi>m</mi></mrow><mrow><mn>2</mn></mrow></msub></math></span> cards of type black (clubs or spades). Given a deck of <span><math><mrow><mi>M</mi><mo>=</mo><msub><mrow><mi>m</mi></mrow><mrow><mn>1</mn></mrow></msub><mo>+</mo><msub><mrow><mi>m</mi></mrow><mrow><mn>2</mn></mrow></msub></mrow></math></span> cards, we propose a refined counting of the number of correct colour guesses, when the guesser is provided with complete information, in other words, when the numbers <span><math><msub><mrow><mi>m</mi></mrow><mrow><mn>1</mn></mrow></msub></math></span> and <span><math><msub><mrow><mi>m</mi></mrow><mrow><mn>2</mn></mrow></msub></math></span> and the colour of each drawn card are known. We decompose the correct guessed cards into three different types by taking into account the probability of making a correct guess, and provide joint distributional results for the underlying random variables as well as joint limit laws.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106160"},"PeriodicalIF":0.9,"publicationDate":"2024-02-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139924736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-10DOI: 10.1016/j.jspi.2024.106153
Li-Pang Chen
Ultrahigh-dimensional data analysis has been a popular topic in decades. In the framework of ultrahigh-dimensional setting, feature screening methods are key techniques to retain informative covariates and screen out non-informative ones when the dimension of covariates is extremely larger than the sample size. In the presence of incomplete data caused by censoring, several valid methods have also been developed to deal with ultrahigh-dimensional covariates for time-to-event data. However, little approach is available to handle feature screening for survival data subject to biased sample, which is usually induced by left-truncation. In this paper, we extend the C-index estimation proposed by Hartman et al. (2023) to develop a valid feature screening procedure to deal with left-truncated and right-censored survival data subject to ultrahigh-dimensional covariates. The sure screening property is also rigorously established to justify the proposed method. Numerical results also verify the validity of the proposed procedure.
几十年来,超高维数据分析一直是一个热门话题。在超高维设置的框架下,当协变量的维度比样本量大得多时,特征筛选方法是保留有信息量的协变量并筛选出无信息量的协变量的关键技术。在普查导致数据不完整的情况下,也开发出了几种有效的方法来处理时间到事件数据的超高维协变量。然而,目前还没有什么方法可以处理生存数据的特征筛选问题,因为生存数据的样本存在偏差,而偏差通常是由左截断引起的。在本文中,我们扩展了 Hartman 等人(2023 年)提出的 C 指数估计方法,开发出一种有效的特征筛选程序,用于处理左截断和右删失的超高维协变量生存数据。此外,还严格建立了确定的筛选属性,以证明所提出的方法是正确的。数值结果也验证了所提方法的有效性。
{"title":"Feature screening via concordance indices for left-truncated and right-censored survival data","authors":"Li-Pang Chen","doi":"10.1016/j.jspi.2024.106153","DOIUrl":"10.1016/j.jspi.2024.106153","url":null,"abstract":"<div><p>Ultrahigh-dimensional data analysis has been a popular topic in decades. In the framework of ultrahigh-dimensional setting, feature screening methods are key techniques to retain informative covariates and screen out non-informative ones when the dimension of covariates is extremely larger than the sample size. In the presence of incomplete data caused by censoring, several valid methods have also been developed to deal with ultrahigh-dimensional covariates for time-to-event data. However, little approach is available to handle feature screening for survival data subject to biased sample, which is usually induced by left-truncation. In this paper, we extend the C-index estimation proposed by Hartman et al. (2023) to develop a valid feature screening procedure to deal with left-truncated and right-censored survival data subject to ultrahigh-dimensional covariates. The sure screening property is also rigorously established to justify the proposed method. Numerical results also verify the validity of the proposed procedure.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106153"},"PeriodicalIF":0.9,"publicationDate":"2024-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139816268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-03DOI: 10.1016/j.jspi.2024.106151
Khouzeima Moutanabbir , Mohammed Bouaddi
In this paper, we address the problem of kernel estimation of the Expected Shortfall (ES) risk measure for financial losses that satisfy the -mixing conditions. First, we introduce a new non-parametric estimator for the ES measure using a kernel estimation. Given that the ES measure is the sum of the Value-at-Risk and the mean-excess function, we provide an estimation of the ES as a sum of the estimators of these two components. Our new estimator has a closed-form expression that depends on the choice of the kernel smoothing function, and we derive these expressions in the case of Gaussian, Uniform, and Epanechnikov kernel functions. We study the asymptotic properties of this new estimator and compare it to the Scaillet estimator. Capitalizing on the properties of these two estimators, we combine them to create a new estimator for the ES which reduces the bias and lowers the mean square error. The combined estimator shows better stability with respect to the choice of the kernel smoothing parameter. Our findings are illustrated through some numerical examples that help us to assess the small sample properties of the different estimators considered in this paper.
本文探讨了满足 α 混合条件的金融损失的预期缺口(ES)风险度量的核估计问题。首先,我们使用核估计法为 ES 度量引入了一个新的非参数估计器。鉴于 ES 度量是风险价值和均值溢出函数之和,我们将 ES 估计为这两个部分的估计值之和。我们的新估计器有一个闭式表达式,它取决于核平滑函数的选择,我们在高斯、均匀和 Epanechnikov 核函数的情况下推导出了这些表达式。我们研究了这种新估计器的渐近特性,并将其与斯凯莱估计器进行了比较。利用这两个估计器的特性,我们将它们结合起来,为 ES 创建了一个新的估计器,从而减少了偏差,降低了均方误差。在选择核平滑参数时,组合估计器显示出更好的稳定性。我们通过一些数字例子来说明我们的发现,这些例子有助于我们评估本文所考虑的不同估计器的小样本特性。
{"title":"A new non-parametric estimation of the expected shortfall for dependent financial losses","authors":"Khouzeima Moutanabbir , Mohammed Bouaddi","doi":"10.1016/j.jspi.2024.106151","DOIUrl":"10.1016/j.jspi.2024.106151","url":null,"abstract":"<div><p>In this paper, we address the problem of kernel estimation of the Expected Shortfall (ES) risk measure for financial losses that satisfy the <span><math><mi>α</mi></math></span>-mixing conditions. First, we introduce a new non-parametric estimator for the ES measure using a kernel estimation. Given that the ES measure is the sum of the Value-at-Risk and the mean-excess function, we provide an estimation of the ES as a sum of the estimators of these two components. Our new estimator has a closed-form expression that depends on the choice of the kernel smoothing function, and we derive these expressions in the case of Gaussian, Uniform, and Epanechnikov kernel functions. We study the asymptotic properties of this new estimator and compare it to the Scaillet estimator. Capitalizing on the properties of these two estimators, we combine them to create a new estimator for the ES which reduces the bias and lowers the mean square error. The combined estimator shows better stability with respect to the choice of the kernel smoothing parameter. Our findings are illustrated through some numerical examples that help us to assess the small sample properties of the different estimators considered in this paper.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106151"},"PeriodicalIF":0.9,"publicationDate":"2024-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0378375824000089/pdfft?md5=41ea07fd0d26fc2bbea00de05c1c0468&pid=1-s2.0-S0378375824000089-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139680115","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-03DOI: 10.1016/j.jspi.2024.106152
Maria Kateri, Nikolay I. Nikolov
Step-stress is a special type of accelerated life-testing procedure that allows the experimenter to test the units of interest under various stress conditions changed (usually increased) at different intermediate time points. In this paper, we study the problem of testing hypothesis for the scale parameter of a simple step-stress model with exponential lifetimes and under Type-II censoring. We consider several modifications of the log-likelihood ratio statistic and eliminate the distributional dependence on the unknown lifetime parameters by exploiting the scale invariant properties of the normalized failure spacings. The presented results and the ratio statistic are further generalized to the multilevel step-stress case under the log-link assumption. We compare the power performance of the proposed tests via Monte Carlo simulations. As an illustration, the described procedures are applied to a real data example from the literature.
阶跃应力是一种特殊的加速寿命测试程序,它允许实验者在不同的中间时间点,在各种应力条件改变(通常是增加)的情况下测试相关单位。在本文中,我们研究了在指数生命期和 II 类删减条件下对简单阶跃应力模型的规模参数进行假设检验的问题。我们考虑了对数似然比统计量的几种修正,并利用归一化失效间隔的尺度不变特性消除了未知寿命参数的分布依赖性。所提出的结果和比值统计量被进一步推广到对数链接假设下的多级阶跃应力情况。我们通过蒙特卡罗模拟比较了所提出的测试的功率性能。作为说明,我们将所述程序应用于文献中的一个真实数据示例。
{"title":"Scale tests for a multilevel step-stress model with exponential lifetimes under Type-II censoring","authors":"Maria Kateri, Nikolay I. Nikolov","doi":"10.1016/j.jspi.2024.106152","DOIUrl":"10.1016/j.jspi.2024.106152","url":null,"abstract":"<div><p>Step-stress is a special type of accelerated life-testing procedure that allows the experimenter to test the units of interest under various stress conditions changed (usually increased) at different intermediate time points. In this paper, we study the problem of testing hypothesis for the scale parameter of a simple step-stress model with exponential lifetimes and under Type-II censoring. We consider several modifications of the log-likelihood ratio statistic and eliminate the distributional dependence on the unknown lifetime parameters by exploiting the scale invariant properties of the normalized failure spacings. The presented results and the ratio statistic are further generalized to the multilevel step-stress case under the log-link assumption. We compare the power performance of the proposed tests via Monte Carlo simulations. As an illustration, the described procedures are applied to a real data example from the literature.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106152"},"PeriodicalIF":0.9,"publicationDate":"2024-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0378375824000090/pdfft?md5=cae47c9c8ceeff2301a8594614cd022f&pid=1-s2.0-S0378375824000090-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139678740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-01DOI: 10.1016/j.jspi.2024.106150
Xu He , Fasheng Sun
Space-filling designs that possess high separation distance are useful for computer experiments. We propose a novel method to construct high-dimensional high-separation distance designs. The construction involves taking the Kronecker product of sub-Hadamard matrices and rotation. In addition to possessing better separation distance than most existing types of space-filling designs, our newly proposed designs enjoy orthogonality and projection uniformity and are more flexible in the numbers of runs and factors than that from most algebraic constructions. From numerical results, such designs are excellent in Gaussian process emulation of high-dimensional computer experiments. An R package on design construction is available online.
具有高分离距离的空间填充设计对计算机实验非常有用。我们提出了一种构建高维高分离距离设计的新方法。这种构建方法涉及子哈达玛矩阵的克朗内克乘积和旋转。与大多数现有的空间填充设计相比,我们新提出的设计除了具有更好的分离距离外,还具有正交性和投影均匀性,并且在运行数和因子数方面比大多数代数构造更加灵活。从数值结果来看,这种设计在高维计算机实验的高斯过程仿真中表现出色。有关设计构造的 R 软件包可在线获取。
{"title":"Construction of high-dimensional high-separation distance designs","authors":"Xu He , Fasheng Sun","doi":"10.1016/j.jspi.2024.106150","DOIUrl":"10.1016/j.jspi.2024.106150","url":null,"abstract":"<div><p>Space-filling designs that possess high separation distance are useful for computer experiments. We propose a novel method to construct high-dimensional high-separation distance designs. The construction involves taking the Kronecker product of sub-Hadamard matrices and rotation. In addition to possessing better separation distance than most existing types of space-filling designs, our newly proposed designs enjoy orthogonality and projection uniformity and are more flexible in the numbers of runs and factors than that from most algebraic constructions. From numerical results, such designs are excellent in Gaussian process emulation of high-dimensional computer experiments. An R package on design construction is available online.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106150"},"PeriodicalIF":0.9,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139663836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-29DOI: 10.1016/j.jspi.2024.106147
Kexuan Li , Fangfang Wang , Ruiqi Liu , Fan Yang , Zuofeng Shang
Ordinary differential equations (ODEs) are widely used to model complex dynamics that arise in biology, chemistry, engineering, finance, physics, etc. Calibration of a complicated ODE system using noisy data is generally challenging. In this paper, we propose a two-stage nonparametric approach to address this problem. We first extract the de-noised data and their higher order derivatives using boundary kernel method, and then feed them into a sparsely connected deep neural network with rectified linear unit (ReLU) activation function. Our method is able to recover the ODE system without being subject to the curse of dimensionality and the complexity of the ODE structure. We have shown that our method is consistent if the ODE possesses a general modular structure with each modular component involving only a few input variables, and the network architecture is properly chosen. Theoretical properties are corroborated by an extensive simulation study that also demonstrates the effectiveness of the proposed method in finite samples. Finally, we use our method to simultaneously characterize the growth rate of COVID-19 cases from the 50 states of the United States.
{"title":"Calibrating multi-dimensional complex ODE from noisy data via deep neural networks","authors":"Kexuan Li , Fangfang Wang , Ruiqi Liu , Fan Yang , Zuofeng Shang","doi":"10.1016/j.jspi.2024.106147","DOIUrl":"10.1016/j.jspi.2024.106147","url":null,"abstract":"<div><p><span>Ordinary differential equations<span> (ODEs) are widely used to model complex dynamics that arise in biology, chemistry, engineering, finance, physics, etc. Calibration of a complicated ODE system using noisy data is generally challenging. In this paper, we propose a two-stage nonparametric approach to address this problem. We first extract the de-noised data and their higher order derivatives using boundary kernel method, and then feed them into a sparsely connected </span></span>deep neural network<span> with rectified linear unit (ReLU) activation function. Our method is able to recover the ODE system without being subject to the curse of dimensionality and the complexity of the ODE structure. We have shown that our method is consistent if the ODE possesses a general modular structure with each modular component involving only a few input variables, and the network architecture is properly chosen. Theoretical properties are corroborated by an extensive simulation study that also demonstrates the effectiveness of the proposed method in finite samples. Finally, we use our method to simultaneously characterize the growth rate of COVID-19 cases from the 50 states of the United States.</span></p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106147"},"PeriodicalIF":0.9,"publicationDate":"2024-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139588540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-26DOI: 10.1016/j.jspi.2024.106149
Jing Zhang , Bo Li , Yu Wang , Xinyi Wei , Xiaohui Liu
In this paper, we suggest an empirical likelihood-based test for the autoregressive coefficient of an integer-valued AR(1) model, i.e., INAR(1). We derive the limit distributions of the resulting test statistic under both null and alternative hypotheses. It turns out that regardless of whether the INAR process is stable or unstable, the statistic is always chi-squared distributed asymptotically under the null hypothesis, and as a result, it can offer unified inferences for the autoregressive coefficient. The performance of its finite sample is also demonstrated using simulations and an empirical example.
{"title":"An empirical likelihood-based unified test for the integer-valued AR(1) models","authors":"Jing Zhang , Bo Li , Yu Wang , Xinyi Wei , Xiaohui Liu","doi":"10.1016/j.jspi.2024.106149","DOIUrl":"10.1016/j.jspi.2024.106149","url":null,"abstract":"<div><p>In this paper, we suggest an empirical likelihood-based test for the autoregressive coefficient of an integer-valued AR(1) model, i.e., INAR(1). We derive the limit distributions of the resulting test statistic under both null and alternative hypotheses. It turns out that regardless of whether the INAR process is stable or unstable, the statistic is always chi-squared distributed asymptotically under the null hypothesis, and as a result, it can offer unified inferences for the autoregressive coefficient. The performance of its finite sample is also demonstrated using simulations and an empirical example.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106149"},"PeriodicalIF":0.9,"publicationDate":"2024-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0378375824000065/pdfft?md5=1c6d378b469788f0758b1d5699e2f871&pid=1-s2.0-S0378375824000065-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139588927","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}