首页 > 最新文献

Statistics and Computing最新文献

英文 中文
Screen then select: a strategy for correlated predictors in high-dimensional quantile regression 先筛选后选择:高维量子回归中相关预测因子的策略
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-04-08 DOI: 10.1007/s11222-024-10424-6
Xuejun Jiang, Yakun Liang, Haofeng Wang

Strong correlation among predictors and heavy-tailed noises pose a great challenge in the analysis of ultra-high dimensional data. Such challenge leads to an increase in the computation time for discovering active variables and a decrease in selection accuracy. To address this issue, we propose an innovative two-stage screen-then-select approach and its derivative procedure based on a robust quantile regression with sparsity assumption. This approach initially screens important features by ranking quantile ridge estimation and subsequently employs a likelihood-based post-screening selection strategy to refine variable selection. Additionally, we conduct an internal competition mechanism along the greedy search path to enhance the robustness of algorithm against the design dependence. Our methods are simple to implement and possess numerous desirable properties from theoretical and computational standpoints. Theoretically, we establish the strong consistency of feature selection for the proposed methods under some regularity conditions. In empirical studies, we assess the finite sample performance of our methods by comparing them with utility screening approaches and existing penalized quantile regression methods. Furthermore, we apply our methods to identify genes associated with anticancer drug sensitivities for practical guidance.

预测因子之间的强相关性和重尾噪声给超高维数据分析带来了巨大挑战。这种挑战导致发现活动变量的计算时间增加,选择精度降低。为了解决这个问题,我们提出了一种创新的 "筛选-选择 "两阶段方法及其基于稀疏性假设的稳健量化回归的衍生程序。这种方法首先通过排序量化脊估计筛选重要特征,然后采用基于似然法的筛选后选择策略来完善变量选择。此外,我们还在贪婪搜索路径上建立了内部竞争机制,以增强算法对设计依赖性的鲁棒性。我们的方法简单易用,而且从理论和计算角度来看具有许多理想特性。从理论上讲,我们建立了所提方法在某些规则性条件下特征选择的强一致性。在实证研究中,我们通过与效用筛选方法和现有的惩罚性量化回归方法进行比较,评估了我们的方法的有限样本性能。此外,我们还将我们的方法用于识别与抗癌药物敏感性相关的基因,以提供实际指导。
{"title":"Screen then select: a strategy for correlated predictors in high-dimensional quantile regression","authors":"Xuejun Jiang, Yakun Liang, Haofeng Wang","doi":"10.1007/s11222-024-10424-6","DOIUrl":"https://doi.org/10.1007/s11222-024-10424-6","url":null,"abstract":"<p>Strong correlation among predictors and heavy-tailed noises pose a great challenge in the analysis of ultra-high dimensional data. Such challenge leads to an increase in the computation time for discovering active variables and a decrease in selection accuracy. To address this issue, we propose an innovative two-stage screen-then-select approach and its derivative procedure based on a robust quantile regression with sparsity assumption. This approach initially screens important features by ranking quantile ridge estimation and subsequently employs a likelihood-based post-screening selection strategy to refine variable selection. Additionally, we conduct an internal competition mechanism along the greedy search path to enhance the robustness of algorithm against the design dependence. Our methods are simple to implement and possess numerous desirable properties from theoretical and computational standpoints. Theoretically, we establish the strong consistency of feature selection for the proposed methods under some regularity conditions. In empirical studies, we assess the finite sample performance of our methods by comparing them with utility screening approaches and existing penalized quantile regression methods. Furthermore, we apply our methods to identify genes associated with anticancer drug sensitivities for practical guidance.</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140573257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
R-VGAL: a sequential variational Bayes algorithm for generalised linear mixed models R-VGAL:广义线性混合模型的顺序变异贝叶斯算法
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-04-06 DOI: 10.1007/s11222-024-10422-8
Bao Anh Vu, David Gunawan, Andrew Zammit-Mangion

Models with random effects, such as generalised linear mixed models (GLMMs), are often used for analysing clustered data. Parameter inference with these models is difficult because of the presence of cluster-specific random effects, which must be integrated out when evaluating the likelihood function. Here, we propose a sequential variational Bayes algorithm, called Recursive Variational Gaussian Approximation for Latent variable models (R-VGAL), for estimating parameters in GLMMs. The R-VGAL algorithm operates on the data sequentially, requires only a single pass through the data, and can provide parameter updates as new data are collected without the need of re-processing the previous data. At each update, the R-VGAL algorithm requires the gradient and Hessian of a “partial” log-likelihood function evaluated at the new observation, which are generally not available in closed form for GLMMs. To circumvent this issue, we propose using an importance-sampling-based approach for estimating the gradient and Hessian via Fisher’s and Louis’ identities. We find that R-VGAL can be unstable when traversing the first few data points, but that this issue can be mitigated by introducing a damping factor in the initial steps of the algorithm. Through illustrations on both simulated and real datasets, we show that R-VGAL provides good approximations to posterior distributions, that it can be made robust through damping, and that it is computationally efficient.

随机效应模型,如广义线性混合模型(GLMM),通常用于分析聚类数据。这些模型的参数推断比较困难,因为存在特定集群的随机效应,在评估似然函数时必须将其整合出来。在此,我们提出了一种序列变异贝叶斯算法,称为潜变量模型的递归变异高斯逼近(R-VGAL),用于估计 GLMM 的参数。R-VGAL 算法按顺序对数据进行操作,只需对数据进行一次传递,并能在收集到新数据时提供参数更新,而无需重新处理以前的数据。每次更新时,R-VGAL 算法都需要在新观测值处评估 "部分 "对数似然函数的梯度和赫塞斯,而 GLMM 通常无法以封闭形式获得这些数据。为了规避这个问题,我们提出了一种基于重要性抽样的方法,通过费雪和路易斯等式来估计梯度和赫斯。我们发现,R-VGAL 在遍历最初几个数据点时可能不稳定,但可以通过在算法初始步骤中引入阻尼因子来缓解这一问题。通过对模拟数据集和真实数据集的举例说明,我们发现 R-VGAL 可以很好地逼近后验分布,通过阻尼可以使其变得稳健,而且计算效率很高。
{"title":"R-VGAL: a sequential variational Bayes algorithm for generalised linear mixed models","authors":"Bao Anh Vu, David Gunawan, Andrew Zammit-Mangion","doi":"10.1007/s11222-024-10422-8","DOIUrl":"https://doi.org/10.1007/s11222-024-10422-8","url":null,"abstract":"<p>Models with random effects, such as generalised linear mixed models (GLMMs), are often used for analysing clustered data. Parameter inference with these models is difficult because of the presence of cluster-specific random effects, which must be integrated out when evaluating the likelihood function. Here, we propose a sequential variational Bayes algorithm, called Recursive Variational Gaussian Approximation for Latent variable models (R-VGAL), for estimating parameters in GLMMs. The R-VGAL algorithm operates on the data sequentially, requires only a single pass through the data, and can provide parameter updates as new data are collected without the need of re-processing the previous data. At each update, the R-VGAL algorithm requires the gradient and Hessian of a “partial” log-likelihood function evaluated at the new observation, which are generally not available in closed form for GLMMs. To circumvent this issue, we propose using an importance-sampling-based approach for estimating the gradient and Hessian via Fisher’s and Louis’ identities. We find that R-VGAL can be unstable when traversing the first few data points, but that this issue can be mitigated by introducing a damping factor in the initial steps of the algorithm. Through illustrations on both simulated and real datasets, we show that R-VGAL provides good approximations to posterior distributions, that it can be made robust through damping, and that it is computationally efficient.</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-04-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140573144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated generation of initial points for adaptive rejection sampling of log-concave distributions 自动生成对数凹分布自适应剔除采样的初始点
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-04-05 DOI: 10.1007/s11222-024-10425-5
Jonathan James

Adaptive rejection sampling requires that users provide points that span the distribution’s mode. If these points are far from the mode, it significantly increases computational costs. This paper introduces a simple, automated approach for selecting initial points that uses numerical optimization to quickly bracket the mode. When an initial point is given that resides in a high-density area, the method often requires just four function evaluations to draw a sample—just one more than the sampler’s minimum. This feature makes it well-suited for Gibbs sampling, where the previous round’s draw can serve as the starting point.

自适应剔除采样要求用户提供跨越分布模式的点。如果这些点远离模式,就会大大增加计算成本。本文介绍了一种简单的自动选择初始点的方法,该方法使用数值优化来快速包围模式。当给定的初始点位于高密度区域时,该方法通常只需进行四次函数求值即可提取样本,仅比采样器的最小值多一次。这一特点使它非常适合吉布斯采样,上一轮的采样可以作为起点。
{"title":"Automated generation of initial points for adaptive rejection sampling of log-concave distributions","authors":"Jonathan James","doi":"10.1007/s11222-024-10425-5","DOIUrl":"https://doi.org/10.1007/s11222-024-10425-5","url":null,"abstract":"<p>Adaptive rejection sampling requires that users provide points that span the distribution’s mode. If these points are far from the mode, it significantly increases computational costs. This paper introduces a simple, automated approach for selecting initial points that uses numerical optimization to quickly bracket the mode. When an initial point is given that resides in a high-density area, the method often requires just four function evaluations to draw a sample—just one more than the sampler’s minimum. This feature makes it well-suited for Gibbs sampling, where the previous round’s draw can serve as the starting point.\u0000</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140573359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Parsimonious ultrametric Gaussian mixture models 解析超参量高斯混合物模型
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-04-01 DOI: 10.1007/s11222-024-10405-9
Carlo Cavicchia, Maurizio Vichi, Giorgia Zaccaria

Gaussian mixture models represent a conceptually and mathematically elegant class of models for casting the density of a heterogeneous population where the observed data is collected from a population composed of a finite set of G homogeneous subpopulations with a Gaussian distribution. A limitation of these models is that they suffer from the curse of dimensionality, and the number of parameters becomes easily extremely large in the presence of high-dimensional data. In this paper, we propose a class of parsimonious Gaussian mixture models with constrained extended ultrametric covariance structures that are capable of exploring hierarchical relations among variables. The proposal shows to require a reduced number of parameters to be fit and includes constrained covariance structures across and within components that further reduce the number of parameters of the model.

高斯混合模型是一类概念上和数学上都很优雅的模型,用于计算异质种群的密度,其中观测数据是从由具有高斯分布的 G 个同质子种群的有限集合组成的种群中收集的。这些模型的局限性在于它们受到维度诅咒的影响,在存在高维数据的情况下,参数数量很容易变得极其庞大。在本文中,我们提出了一类具有受限扩展超对称协方差结构的简约高斯混合物模型,这些模型能够探索变量之间的层次关系。该建议表明,拟合所需的参数数量减少了,并且包括跨成分和成分内部的约束协方差结构,从而进一步减少了模型的参数数量。
{"title":"Parsimonious ultrametric Gaussian mixture models","authors":"Carlo Cavicchia, Maurizio Vichi, Giorgia Zaccaria","doi":"10.1007/s11222-024-10405-9","DOIUrl":"https://doi.org/10.1007/s11222-024-10405-9","url":null,"abstract":"<p>Gaussian mixture models represent a conceptually and mathematically elegant class of models for casting the density of a heterogeneous population where the observed data is collected from a population composed of a finite set of <i>G</i> homogeneous subpopulations with a Gaussian distribution. A limitation of these models is that they suffer from the curse of dimensionality, and the number of parameters becomes easily extremely large in the presence of high-dimensional data. In this paper, we propose a class of parsimonious Gaussian mixture models with constrained extended ultrametric covariance structures that are capable of exploring hierarchical relations among variables. The proposal shows to require a reduced number of parameters to be fit and includes constrained covariance structures across and within components that further reduce the number of parameters of the model.</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140573121","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stochastic three-term conjugate gradient method with variance technique for non-convex learning 用于非凸学习的随机三项共轭梯度法与方差技术
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-03-27 DOI: 10.1007/s11222-024-10409-5
Chen Ouyang, Chenkaixiang Lu, Xiong Zhao, Ruping Huang, Gonglin Yuan, Yiyan Jiang

In the training process of machine learning, the minimization of the empirical risk loss function is often used to measure the difference between the model’s predicted value and the real value. Stochastic gradient descent is very popular for this type of optimization problem, but converges slowly in theoretical analysis. To solve this problem, there are already many algorithms with variance reduction techniques, such as SVRG, SAG, SAGA, etc. Some scholars apply the conjugate gradient method in traditional optimization to these algorithms, such as CGVR, SCGA, SCGN, etc., which can basically achieve linear convergence speed, but these conclusions often need to be established under some relatively strong assumptions. In traditional optimization, the conjugate gradient method often requires the use of line search techniques to achieve good experimental results. In a sense, line search embodies some properties of the conjugate methods. Taking inspiration from this, we apply the modified three-term conjugate gradient method and line search technique to machine learning. In our theoretical analysis, we obtain the same convergence rate as SCGA under weaker conditional assumptions. We also test the convergence of our algorithm using two non-convex machine learning models.

在机器学习的训练过程中,经验风险损失函数的最小化通常用于衡量模型预测值与实际值之间的差异。随机梯度下降法在这类优化问题中非常流行,但从理论分析来看收敛速度较慢。为了解决这个问题,已经有很多算法采用了方差缩小技术,如 SVRG、SAG、SAGA 等。一些学者将传统优化中的共轭梯度法应用到这些算法中,如 CGVR、SCGA、SCGN 等,基本可以达到线性收敛速度,但这些结论往往需要在一些比较强的假设条件下才能成立。在传统优化中,共轭梯度法往往需要使用直线搜索技术才能取得良好的实验结果。从某种意义上说,直线搜索体现了共轭方法的某些特性。受此启发,我们将改进的三期共轭梯度法和线搜索技术应用于机器学习。在理论分析中,我们在较弱的条件假设下获得了与 SCGA 相同的收敛率。我们还利用两个非凸机器学习模型测试了算法的收敛性。
{"title":"Stochastic three-term conjugate gradient method with variance technique for non-convex learning","authors":"Chen Ouyang, Chenkaixiang Lu, Xiong Zhao, Ruping Huang, Gonglin Yuan, Yiyan Jiang","doi":"10.1007/s11222-024-10409-5","DOIUrl":"https://doi.org/10.1007/s11222-024-10409-5","url":null,"abstract":"<p>In the training process of machine learning, the minimization of the empirical risk loss function is often used to measure the difference between the model’s predicted value and the real value. Stochastic gradient descent is very popular for this type of optimization problem, but converges slowly in theoretical analysis. To solve this problem, there are already many algorithms with variance reduction techniques, such as SVRG, SAG, SAGA, etc. Some scholars apply the conjugate gradient method in traditional optimization to these algorithms, such as CGVR, SCGA, SCGN, etc., which can basically achieve linear convergence speed, but these conclusions often need to be established under some relatively strong assumptions. In traditional optimization, the conjugate gradient method often requires the use of line search techniques to achieve good experimental results. In a sense, line search embodies some properties of the conjugate methods. Taking inspiration from this, we apply the modified three-term conjugate gradient method and line search technique to machine learning. In our theoretical analysis, we obtain the same convergence rate as SCGA under weaker conditional assumptions. We also test the convergence of our algorithm using two non-convex machine learning models.</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140311638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Novel sampling method for the von Mises–Fisher distribution von Mises-Fisher 分布的新型抽样方法
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-03-26 DOI: 10.1007/s11222-024-10419-3

Abstract

The von Mises–Fisher distribution is a widely used probability model in directional statistics. An algorithm for generating pseudo-random vectors from this distribution was suggested by Wood (Commun Stat Simul Comput 23(1):157–164, 1994), which is based on a rejection sampling scheme. This paper proposes an alternative to this rejection sampling approach for drawing pseudo-random vectors from arbitrary von Mises–Fisher distributions. A useful mixture representation is derived, which is a mixture of beta distributions with mixing weights that follow a confluent hypergeometric distribution. A condensed table-lookup method is adopted for sampling from the confluent hypergeometric distribution. A theoretical analysis investigates the amount of computation required to construct the condensed lookup table. Through numerical experiments, we demonstrate that the proposed algorithm outperforms the rejection-based method when generating a large number of pseudo-random vectors from the same distribution.

摘要 von Mises-Fisher 分布是定向统计中广泛使用的概率模型。伍德(Commun Stat Simul Comput 23(1):157-164, 1994)提出了一种从该分布生成伪随机向量的算法,该算法基于拒绝抽样方案。本文提出了从任意 von Mises-Fisher 分布中抽取伪随机向量的拒绝抽样方法的替代方案。本文导出了一种有用的混合表示法,即混合权重遵循汇合超几何分布的贝塔分布的混合。从汇合超几何分布中采样时,采用了一种浓缩的查表方法。理论分析研究了构建浓缩查找表所需的计算量。通过数值实验,我们证明了当从同一分布生成大量伪随机向量时,所提出的算法优于基于拒绝的方法。
{"title":"Novel sampling method for the von Mises–Fisher distribution","authors":"","doi":"10.1007/s11222-024-10419-3","DOIUrl":"https://doi.org/10.1007/s11222-024-10419-3","url":null,"abstract":"<h3>Abstract</h3> <p>The von Mises–Fisher distribution is a widely used probability model in directional statistics. An algorithm for generating pseudo-random vectors from this distribution was suggested by Wood (Commun Stat Simul Comput 23(1):157–164, 1994), which is based on a rejection sampling scheme. This paper proposes an alternative to this rejection sampling approach for drawing pseudo-random vectors from arbitrary von Mises–Fisher distributions. A useful mixture representation is derived, which is a mixture of beta distributions with mixing weights that follow a confluent hypergeometric distribution. A condensed table-lookup method is adopted for sampling from the confluent hypergeometric distribution. A theoretical analysis investigates the amount of computation required to construct the condensed lookup table. Through numerical experiments, we demonstrate that the proposed algorithm outperforms the rejection-based method when generating a large number of pseudo-random vectors from the same distribution.</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140301529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalized spherical principal component analysis 广义球形主成分分析
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-03-23 DOI: 10.1007/s11222-024-10413-9
Sarah Leyder, Jakob Raymaekers, Tim Verdonck

Outliers contaminating data sets are a challenge to statistical estimators. Even a small fraction of outlying observations can heavily influence most classical statistical methods. In this paper we propose generalized spherical principal component analysis, a new robust version of principal component analysis that is based on the generalized spatial sign covariance matrix. Theoretical properties of the proposed method including influence functions, breakdown values and asymptotic efficiencies are derived. These theoretical results are complemented with an extensive simulation study and two real-data examples. We illustrate that generalized spherical principal component analysis can combine great robustness with solid efficiency properties, in addition to a low computational cost.

污染数据集的异常值是对统计估计方法的挑战。即使是一小部分离群观测值,也会严重影响大多数经典统计方法。在本文中,我们提出了广义球形主成分分析法,这是一种基于广义空间符号协方差矩阵的新的稳健型主成分分析法。本文推导了所提方法的理论特性,包括影响函数、崩溃值和渐近效率。这些理论结果通过广泛的模拟研究和两个真实数据实例得到了补充。我们说明,除了计算成本低廉之外,广义球形主成分分析法还能将强大的鲁棒性与可靠的效率特性结合起来。
{"title":"Generalized spherical principal component analysis","authors":"Sarah Leyder, Jakob Raymaekers, Tim Verdonck","doi":"10.1007/s11222-024-10413-9","DOIUrl":"https://doi.org/10.1007/s11222-024-10413-9","url":null,"abstract":"<p>Outliers contaminating data sets are a challenge to statistical estimators. Even a small fraction of outlying observations can heavily influence most classical statistical methods. In this paper we propose generalized spherical principal component analysis, a new robust version of principal component analysis that is based on the generalized spatial sign covariance matrix. Theoretical properties of the proposed method including influence functions, breakdown values and asymptotic efficiencies are derived. These theoretical results are complemented with an extensive simulation study and two real-data examples. We illustrate that generalized spherical principal component analysis can combine great robustness with solid efficiency properties, in addition to a low computational cost.</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140201283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Variable selection using axis-aligned random projections for partial least-squares regression 利用轴对齐随机投影为部分最小二乘回归选择变量
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-03-23 DOI: 10.1007/s11222-024-10417-5

Abstract

In high-dimensional data modeling, variable selection plays a crucial role in improving predictive accuracy and enhancing model interpretability through sparse representation. Unfortunately, certain variable selection methods encounter challenges such as insufficient model sparsity, high computational overhead, and difficulties in handling large-scale data. Recently, axis-aligned random projection techniques have been applied to address these issues by selecting variables. However, these techniques have seen limited application in handling complex data within the regression framework. In this study, we propose a novel method, sparse partial least squares via axis-aligned random projection, designed for the analysis of high-dimensional data. Initially, axis-aligned random projection is utilized to obtain a sparse loading vector, significantly reducing computational complexity. Subsequently, partial least squares regression is conducted within the subspace of the top-ranked significant variables. The submatrices are iteratively updated until an optimal sparse partial least squares model is achieved. Comparative analysis with some state-of-the-art high-dimensional regression methods demonstrates that the proposed method exhibits superior predictive performance. To illustrate its effectiveness, we apply the method to four cases, including one simulated dataset and three real-world datasets. The results show the proposed method’s ability to identify important variables in all four cases.

摘要 在高维数据建模中,变量选择在通过稀疏表示提高预测精度和增强模型可解释性方面起着至关重要的作用。遗憾的是,某些变量选择方法面临着模型稀疏性不足、计算开销大、难以处理大规模数据等挑战。最近,轴对齐随机投影技术被用于通过选择变量来解决这些问题。然而,这些技术在回归框架内处理复杂数据时应用有限。在本研究中,我们提出了一种新方法--通过轴对齐随机投影的稀疏偏最小二乘法,专门用于分析高维数据。首先,利用轴对齐随机投影获得稀疏载荷向量,大大降低了计算复杂度。随后,在排名靠前的重要变量子空间内进行偏最小二乘法回归。对子矩阵进行迭代更新,直到获得最佳稀疏偏最小二乘法模型。与一些最先进的高维回归方法的比较分析表明,所提出的方法具有卓越的预测性能。为了说明该方法的有效性,我们在四个案例中应用了该方法,包括一个模拟数据集和三个实际数据集。结果表明,所提出的方法能够在所有四种情况下识别重要变量。
{"title":"Variable selection using axis-aligned random projections for partial least-squares regression","authors":"","doi":"10.1007/s11222-024-10417-5","DOIUrl":"https://doi.org/10.1007/s11222-024-10417-5","url":null,"abstract":"<h3>Abstract</h3> <p>In high-dimensional data modeling, variable selection plays a crucial role in improving predictive accuracy and enhancing model interpretability through sparse representation. Unfortunately, certain variable selection methods encounter challenges such as insufficient model sparsity, high computational overhead, and difficulties in handling large-scale data. Recently, axis-aligned random projection techniques have been applied to address these issues by selecting variables. However, these techniques have seen limited application in handling complex data within the regression framework. In this study, we propose a novel method, sparse partial least squares via axis-aligned random projection, designed for the analysis of high-dimensional data. Initially, axis-aligned random projection is utilized to obtain a sparse loading vector, significantly reducing computational complexity. Subsequently, partial least squares regression is conducted within the subspace of the top-ranked significant variables. The submatrices are iteratively updated until an optimal sparse partial least squares model is achieved. Comparative analysis with some state-of-the-art high-dimensional regression methods demonstrates that the proposed method exhibits superior predictive performance. To illustrate its effectiveness, we apply the method to four cases, including one simulated dataset and three real-world datasets. The results show the proposed method’s ability to identify important variables in all four cases.</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140201284","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An expectile computation cookbook 预期计算食谱
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-03-23 DOI: 10.1007/s11222-024-10403-x

Abstract

A substantial body of work in the last 15 years has shown that expectiles constitute an excellent candidate for becoming a standard tool in probabilistic and statistical modeling. Surprisingly, the question of how expectiles may be efficiently calculated has been left largely untouched. We fill this gap by, first, providing a general outlook on the computation of expectiles that relies on the knowledge of analytic expressions of the underlying distribution function and mean residual life function. We distinguish between discrete distributions, for which an exact calculation is always feasible, and continuous distributions, where a Newton–Raphson approximation algorithm can be implemented and a list of exceptional distributions whose expectiles are in analytic form can be given. When the distribution function and/or the mean residual life is difficult to compute, Monte-Carlo algorithms are introduced, based on an exact calculation of sample expectiles and on the use of control variates to improve computational efficiency. We discuss the relevance of our findings to statistical practice and provide numerical evidence of the performance of the considered methods.

摘要 过去 15 年的大量工作表明,期望值是成为概率和统计建模标准工具的绝佳候选。令人惊讶的是,关于如何有效计算期望值的问题却基本上没有涉及。为了填补这一空白,我们首先提供了关于计算期望值的一般展望,这种展望依赖于对基本分布函数和平均残差生命函数解析表达式的了解。我们将离散型分布和连续型分布区分开来,离散型分布的精确计算总是可行的,而连续型分布则可以采用牛顿-拉斐森近似算法,并给出其期望值为解析形式的特殊分布列表。当分布函数和/或平均残差寿命难以计算时,我们引入了蒙特卡洛算法,该算法基于对样本期望值的精确计算,并使用控制变量来提高计算效率。我们讨论了研究结果与统计实践的相关性,并提供了所考虑方法性能的数值证据。
{"title":"An expectile computation cookbook","authors":"","doi":"10.1007/s11222-024-10403-x","DOIUrl":"https://doi.org/10.1007/s11222-024-10403-x","url":null,"abstract":"<h3>Abstract</h3> <p>A substantial body of work in the last 15 years has shown that expectiles constitute an excellent candidate for becoming a standard tool in probabilistic and statistical modeling. Surprisingly, the question of how expectiles may be efficiently calculated has been left largely untouched. We fill this gap by, first, providing a general outlook on the computation of expectiles that relies on the knowledge of analytic expressions of the underlying distribution function and mean residual life function. We distinguish between discrete distributions, for which an exact calculation is always feasible, and continuous distributions, where a Newton–Raphson approximation algorithm can be implemented and a list of exceptional distributions whose expectiles are in analytic form can be given. When the distribution function and/or the mean residual life is difficult to compute, Monte-Carlo algorithms are introduced, based on an exact calculation of sample expectiles and on the use of control variates to improve computational efficiency. We discuss the relevance of our findings to statistical practice and provide numerical evidence of the performance of the considered methods.</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140205561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Simultaneous estimation and variable selection for a non-crossing multiple quantile regression using deep neural networks 利用深度神经网络实现非交叉多元量级回归的同步估计和变量选择
IF 2.2 2区 数学 Q1 Mathematics Pub Date : 2024-03-22 DOI: 10.1007/s11222-024-10418-4
Jungmin Shin, Seunghyun Gwak, Seung Jun Shin, Sungwan Bang

In this paper, we present the DNN-NMQR estimator, an approach that utilizes a deep neural network structure to solve multiple quantile regression problems. When estimating multiple quantiles, our approach leverages the structural characteristics of DNN to enhance estimation results by encouraging shared learning across different quantiles through DNN-NMQR. Also, this method effectively addresses quantile crossing issues through the penalization method. To refine our methodology, we introduce a convolution-type quadratic smoothing function, ensuring that the objective function remains differentiable throughout. Furthermore, we provide a brief discussion on the convergence analysis of DNN-NMQR, drawing on the concept of the neural tangent kernel. For a high-dimensional case, we propose the (A)GDNN-NMQR estimator, which applies group-wise (L_1)-type regularization methods and enjoys the advantages of quantile estimation and variable selection simultaneously. We extensively validate all of our proposed methods through numerical experiments and real data analysis.

本文介绍了 DNN-NMQR 估计器,这是一种利用深度神经网络结构解决多重量化回归问题的方法。在估计多个量化值时,我们的方法利用 DNN 的结构特征,通过 DNN-NMQR 鼓励不同量化值之间的共享学习,从而提高估计结果。此外,该方法还通过惩罚方法有效解决了量值交叉问题。为了完善我们的方法,我们引入了卷积型二次平滑函数,确保目标函数始终保持可微分。此外,我们还借鉴神经正切核的概念,简要讨论了 DNN-NMQR 的收敛分析。对于高维情况,我们提出了 (A)GDNN-NMQR 估计器,该估计器应用了群智(L_1)型正则化方法,同时享有量化估计和变量选择的优势。我们通过数值实验和实际数据分析广泛验证了我们提出的所有方法。
{"title":"Simultaneous estimation and variable selection for a non-crossing multiple quantile regression using deep neural networks","authors":"Jungmin Shin, Seunghyun Gwak, Seung Jun Shin, Sungwan Bang","doi":"10.1007/s11222-024-10418-4","DOIUrl":"https://doi.org/10.1007/s11222-024-10418-4","url":null,"abstract":"<p>In this paper, we present the DNN-NMQR estimator, an approach that utilizes a deep neural network structure to solve multiple quantile regression problems. When estimating multiple quantiles, our approach leverages the structural characteristics of DNN to enhance estimation results by encouraging shared learning across different quantiles through DNN-NMQR. Also, this method effectively addresses quantile crossing issues through the penalization method. To refine our methodology, we introduce a convolution-type quadratic smoothing function, ensuring that the objective function remains differentiable throughout. Furthermore, we provide a brief discussion on the convergence analysis of DNN-NMQR, drawing on the concept of the neural tangent kernel. For a high-dimensional case, we propose the (A)GDNN-NMQR estimator, which applies group-wise <span>(L_1)</span>-type regularization methods and enjoys the advantages of quantile estimation and variable selection simultaneously. We extensively validate all of our proposed methods through numerical experiments and real data analysis.</p>","PeriodicalId":22058,"journal":{"name":"Statistics and Computing","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140201225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Statistics and Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1