首页 > 最新文献

arXiv - STAT - Computation最新文献

英文 中文
Scalable expectation propagation for generalized linear models 广义线性模型的可扩展期望传播
Pub Date : 2024-07-02 DOI: arxiv-2407.02128
Niccolò Anceschi, Augusto Fasano, Beatrice Franzolini, Giovanni Rebaudo
Generalized linear models (GLMs) arguably represent the standard approach forstatistical regression beyond the Gaussian likelihood scenario. When Bayesianformulations are employed, the general absence of a tractable posteriordistribution has motivated the development of deterministic approximations,which are generally more scalable than sampling techniques. Among them,expectation propagation (EP) showed extreme accuracy, usually higher than manyvariational Bayes solutions. However, the higher computational cost of EP posedconcerns about its practical feasibility, especially in high-dimensionalsettings. We address these concerns by deriving a novel efficient formulationof EP for GLMs, whose cost scales linearly in the number of covariates p. Thisreduces the state-of-the-art O(p^2 n) per-iteration computational cost of theEP routine for GLMs to O(p n min{p,n}), with n being the sample size. We alsoshow that, for binary models and log-linear GLMs approximate predictive meanscan be obtained at no additional cost. To preserve efficient moment matchingfor count data, we propose employing a combination of log-normal Laplacetransform approximations, avoiding numerical integration. These novel resultsopen the possibility of employing EP in settings that were believed to bepractically impossible. Improvements over state-of-the-art approaches areillustrated both for simulated and real data. The efficient EP implementationis available at https://github.com/niccoloanceschi/EPglm.
广义线性模型(GLM)可以说是超越高斯似然情景的标准统计回归方法。在使用贝叶斯公式时,由于普遍缺乏可操作的后分布,因此人们开发了确定性近似方法,这种方法通常比抽样技术更具可扩展性。其中,期望传播(EP)显示出极高的准确性,通常高于许多变量贝叶斯解决方案。然而,EP 较高的计算成本使人们对其实际可行性产生了担忧,尤其是在高维环境中。为了解决这些问题,我们为 GLMs 推导了一种新的高效 EP 方案,其成本与协方差的数量 p 成线性比例,从而将 GLMs 的 EP 例程的最新 O(p^2 n) 每次迭代计算成本降至 O(p,n),n 为样本大小。我们还证明,对于二元模型和对数线性 GLM,可以在不增加成本的情况下获得近似预测均值。为了保持计数数据的有效矩匹配,我们建议采用对数正态拉普变换近似的组合,避免数值积分。这些新颖的结果为在人们认为实际上不可能的情况下使用 EP 提供了可能性。在模拟数据和真实数据方面,与最先进的方法相比都有很大改进。高效的 EP 实现可在 https://github.com/niccoloanceschi/EPglm 上获取。
{"title":"Scalable expectation propagation for generalized linear models","authors":"Niccolò Anceschi, Augusto Fasano, Beatrice Franzolini, Giovanni Rebaudo","doi":"arxiv-2407.02128","DOIUrl":"https://doi.org/arxiv-2407.02128","url":null,"abstract":"Generalized linear models (GLMs) arguably represent the standard approach for\u0000statistical regression beyond the Gaussian likelihood scenario. When Bayesian\u0000formulations are employed, the general absence of a tractable posterior\u0000distribution has motivated the development of deterministic approximations,\u0000which are generally more scalable than sampling techniques. Among them,\u0000expectation propagation (EP) showed extreme accuracy, usually higher than many\u0000variational Bayes solutions. However, the higher computational cost of EP posed\u0000concerns about its practical feasibility, especially in high-dimensional\u0000settings. We address these concerns by deriving a novel efficient formulation\u0000of EP for GLMs, whose cost scales linearly in the number of covariates p. This\u0000reduces the state-of-the-art O(p^2 n) per-iteration computational cost of the\u0000EP routine for GLMs to O(p n min{p,n}), with n being the sample size. We also\u0000show that, for binary models and log-linear GLMs approximate predictive means\u0000can be obtained at no additional cost. To preserve efficient moment matching\u0000for count data, we propose employing a combination of log-normal Laplace\u0000transform approximations, avoiding numerical integration. These novel results\u0000open the possibility of employing EP in settings that were believed to be\u0000practically impossible. Improvements over state-of-the-art approaches are\u0000illustrated both for simulated and real data. The efficient EP implementation\u0000is available at https://github.com/niccoloanceschi/EPglm.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141531894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A General Purpose Approximation to the Ferguson-Klass Algorithm for Sampling from Lévy Processes Without Gaussian Components 从无高斯成分的莱维过程采样的弗格森-克拉斯算法的通用近似值
Pub Date : 2024-07-01 DOI: arxiv-2407.01483
Dawid Bernaciak, Jim E. Griffin
We propose a general-purpose approximation to the Ferguson-Klass algorithmfor generating samples from L'evy processes without Gaussian components. Weshow that the proposed method is more than 1000 times faster than the standardFerguson-Klass algorithm without a significant loss of precision. This methodcan open an avenue for computationally efficient and scalable Bayesiannonparametric models which go beyond conjugacy assumptions, as demonstrated inthe examples section.
我们提出了一种通用的近似弗格森-克拉斯算法,用于从没有高斯成分的 L'evy 过程中生成样本。结果表明,所提出的方法比标准的弗格森-克拉斯算法快 1000 多倍,而且精度没有明显下降。正如示例部分所展示的,这种方法可以为计算高效、可扩展的贝叶斯非参数模型开辟一条途径,这些模型超越了共轭假设。
{"title":"A General Purpose Approximation to the Ferguson-Klass Algorithm for Sampling from Lévy Processes Without Gaussian Components","authors":"Dawid Bernaciak, Jim E. Griffin","doi":"arxiv-2407.01483","DOIUrl":"https://doi.org/arxiv-2407.01483","url":null,"abstract":"We propose a general-purpose approximation to the Ferguson-Klass algorithm\u0000for generating samples from L'evy processes without Gaussian components. We\u0000show that the proposed method is more than 1000 times faster than the standard\u0000Ferguson-Klass algorithm without a significant loss of precision. This method\u0000can open an avenue for computationally efficient and scalable Bayesian\u0000nonparametric models which go beyond conjugacy assumptions, as demonstrated in\u0000the examples section.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structured Sketching for Linear Systems 线性系统结构草图
Pub Date : 2024-06-30 DOI: arxiv-2407.00746
Johannes J Brust, Michael A Saunders
For linear systems $Ax=b$ we develop iterative algorithms based on asketch-and-project approach. By using judicious choices for the sketch, such asthe history of residuals, we develop weighting strategies that enable shortrecursive formulas. The proposed algorithms have a low memory footprint anditeration complexity compared to regular sketch-and-project methods. In a setof numerical experiments the new methods compare well to GMRES, SYMMLQ andstate-of-the-art randomized solvers.
对于线性系统 $Ax=b$,我们开发了基于 "求取和项目 "方法的迭代算法。通过对草图(如残差的历史)的明智选择,我们开发出了加权策略,从而实现了短递归公式。与常规的草图-项目法相比,所提出的算法具有较低的内存占用和运算复杂度。在一组数值实验中,新方法与 GMRES、SYMMLQ 和最先进的随机求解器相比都有很好的效果。
{"title":"Structured Sketching for Linear Systems","authors":"Johannes J Brust, Michael A Saunders","doi":"arxiv-2407.00746","DOIUrl":"https://doi.org/arxiv-2407.00746","url":null,"abstract":"For linear systems $Ax=b$ we develop iterative algorithms based on a\u0000sketch-and-project approach. By using judicious choices for the sketch, such as\u0000the history of residuals, we develop weighting strategies that enable short\u0000recursive formulas. The proposed algorithms have a low memory footprint and\u0000iteration complexity compared to regular sketch-and-project methods. In a set\u0000of numerical experiments the new methods compare well to GMRES, SYMMLQ and\u0000state-of-the-art randomized solvers.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141531895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deterministic and Stochastic Frank-Wolfe Recursion on Probability Spaces 概率空间上的确定性和随机性弗兰克-沃尔夫递推
Pub Date : 2024-06-29 DOI: arxiv-2407.00307
Di Yu, Shane G. Henderson, Raghu Pasupathy
Motivated by applications in emergency response and experimental design, weconsider smooth stochastic optimization problems over probability measuressupported on compact subsets of the Euclidean space. With the influencefunction as the variational object, we construct a deterministic Frank-Wolfe(dFW) recursion for probability spaces, made especially possible by a lemmathat identifies a ``closed-form'' solution to the infinite-dimensionalFrank-Wolfe sub-problem. Each iterate in dFW is expressed as a convexcombination of the incumbent iterate and a Dirac measure concentrating on theminimum of the influence function at the incumbent iterate. To address commonapplication contexts that have access only to Monte Carlo observations of theobjective and influence function, we construct a stochastic Frank-Wolfe (sFW)variation that generates a random sequence of probability measures constructedusing minima of increasingly accurate estimates of the influence function. Wedemonstrate that sFW's optimality gap sequence exhibits $O(k^{-1})$ iterationcomplexity almost surely and in expectation for smooth convex objectives, and$O(k^{-1/2})$ (in Frank-Wolfe gap) for smooth non-convex objectives.Furthermore, we show that an easy-to-implement fixed-step, fixed-sample versionof (sFW) exhibits exponential convergence to $varepsilon$-optimality. We endwith a central limit theorem on the observed objective values at the sequenceof generated random measures. To further intuition, we include severalillustrative examples with exact influence function calculations.
受应急响应和实验设计应用的启发,我们考虑了欧几里得空间紧凑子集上支持的概率度量的平稳随机优化问题。以影响函数作为变分对象,我们构建了概率空间的确定性弗兰克-沃尔夫(dFW)递归,特别是通过一个确定无穷维弗兰克-沃尔夫子问题的 "封闭形式 "解的定理,使之成为可能。dFW 中的每个迭代都表示为现任迭代的凸组合和集中于现任迭代处影响函数最小值的狄拉克度量。为了解决只能获得目标和影响函数的蒙特卡罗观测结果的常见应用问题,我们构建了一种随机弗兰克-沃尔夫(sFW)变量,它能生成一系列随机概率度量,这些概率度量是利用对影响函数越来越精确的估计的最小值构建的。我们证明,对于平滑凸目标,sFW 的最优性差距序列几乎肯定地在期望值上表现出 $O(k^{-1})$ 的迭代复杂性,而对于平滑非凸目标,则表现出 $O(k^{-1/2})$ 的迭代复杂性(在 Frank-Wolfe 差距中)。此外,我们还证明,一个易于实现的固定步长、固定样本版本的 (sFW) 表现出指数级收敛到 $varepsilon$ 的最优性。最后,我们提出了一个关于在生成的随机度量序列中观察到的目标值的中心极限定理。为了进一步加深直觉,我们列举了几个具有精确影响函数计算的示例。
{"title":"Deterministic and Stochastic Frank-Wolfe Recursion on Probability Spaces","authors":"Di Yu, Shane G. Henderson, Raghu Pasupathy","doi":"arxiv-2407.00307","DOIUrl":"https://doi.org/arxiv-2407.00307","url":null,"abstract":"Motivated by applications in emergency response and experimental design, we\u0000consider smooth stochastic optimization problems over probability measures\u0000supported on compact subsets of the Euclidean space. With the influence\u0000function as the variational object, we construct a deterministic Frank-Wolfe\u0000(dFW) recursion for probability spaces, made especially possible by a lemma\u0000that identifies a ``closed-form'' solution to the infinite-dimensional\u0000Frank-Wolfe sub-problem. Each iterate in dFW is expressed as a convex\u0000combination of the incumbent iterate and a Dirac measure concentrating on the\u0000minimum of the influence function at the incumbent iterate. To address common\u0000application contexts that have access only to Monte Carlo observations of the\u0000objective and influence function, we construct a stochastic Frank-Wolfe (sFW)\u0000variation that generates a random sequence of probability measures constructed\u0000using minima of increasingly accurate estimates of the influence function. We\u0000demonstrate that sFW's optimality gap sequence exhibits $O(k^{-1})$ iteration\u0000complexity almost surely and in expectation for smooth convex objectives, and\u0000$O(k^{-1/2})$ (in Frank-Wolfe gap) for smooth non-convex objectives.\u0000Furthermore, we show that an easy-to-implement fixed-step, fixed-sample version\u0000of (sFW) exhibits exponential convergence to $varepsilon$-optimality. We end\u0000with a central limit theorem on the observed objective values at the sequence\u0000of generated random measures. To further intuition, we include several\u0000illustrative examples with exact influence function calculations.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141531754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A note on the relationship between PDE-based precision operators and Matérn covariances 关于基于 PDE 的精确算子与马特恩协方差之间关系的说明
Pub Date : 2024-06-29 DOI: arxiv-2407.00471
Umberto Villa, Thomas O'Leary-Roseberry
The purpose of this technical note is to summarize the relationship betweenthe marginal variance and correlation length of a Gaussian random field withMat'ern covariance and the coefficients of the correspondingpartial-differential-equation (PDE)-based precision operator.
本技术说明旨在总结具有马特协方差的高斯随机场的边际方差和相关长度与相应的基于偏微分方程(PDE)的精确算子的系数之间的关系。
{"title":"A note on the relationship between PDE-based precision operators and Matérn covariances","authors":"Umberto Villa, Thomas O'Leary-Roseberry","doi":"arxiv-2407.00471","DOIUrl":"https://doi.org/arxiv-2407.00471","url":null,"abstract":"The purpose of this technical note is to summarize the relationship between\u0000the marginal variance and correlation length of a Gaussian random field with\u0000Mat'ern covariance and the coefficients of the corresponding\u0000partial-differential-equation (PDE)-based precision operator.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141523198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exact confidence intervals for functions of parameters in the k-sample multinomial problem k 样本多项式问题中参数函数的精确置信区间
Pub Date : 2024-06-27 DOI: arxiv-2406.19141
Michael C Sachs, Erin E Gabriel, Michael P Fay
When the target of inference is a real-valued function of probabilityparameters in the k-sample multinomial problem, variance estimation may bechallenging. In small samples, methods like the nonparametric bootstrap ordelta method may perform poorly. We propose a novel general method in thissetting for computing exact p-values and confidence intervals which means thattype I error rates are correctly bounded and confidence intervals have at leastnominal coverage at all sample sizes. Our method is applicable to anyreal-valued function of multinomial probabilities, accommodating an arbitrarynumber of samples with varying category counts. We describe the method andprovide an implementation of it in R, with some computational optimization toensure broad applicability. Simulations demonstrate our method's ability tomaintain correct coverage rates in settings where the nonparametric bootstrapfails.
当推断的目标是 k 样本多项式问题中概率参数的实值函数时,方差估计可能会很困难。在小样本中,像非参数自举阶梯法这样的方法可能会表现不佳。在这种情况下,我们提出了一种计算精确 p 值和置信区间的新颖通用方法,这意味着在所有样本大小下,I 型误差率都能得到正确的约束,置信区间至少有名义覆盖率。我们的方法适用于多项式概率的任何实值函数,可容纳任意数量的具有不同类别计数的样本。我们描述了该方法,并提供了它在 R 语言中的实现,同时进行了一些计算优化,以确保广泛的适用性。模拟证明了我们的方法能够在非参数引导法失效的情况下保持正确的覆盖率。
{"title":"Exact confidence intervals for functions of parameters in the k-sample multinomial problem","authors":"Michael C Sachs, Erin E Gabriel, Michael P Fay","doi":"arxiv-2406.19141","DOIUrl":"https://doi.org/arxiv-2406.19141","url":null,"abstract":"When the target of inference is a real-valued function of probability\u0000parameters in the k-sample multinomial problem, variance estimation may be\u0000challenging. In small samples, methods like the nonparametric bootstrap or\u0000delta method may perform poorly. We propose a novel general method in this\u0000setting for computing exact p-values and confidence intervals which means that\u0000type I error rates are correctly bounded and confidence intervals have at least\u0000nominal coverage at all sample sizes. Our method is applicable to any\u0000real-valued function of multinomial probabilities, accommodating an arbitrary\u0000number of samples with varying category counts. We describe the method and\u0000provide an implementation of it in R, with some computational optimization to\u0000ensure broad applicability. Simulations demonstrate our method's ability to\u0000maintain correct coverage rates in settings where the nonparametric bootstrap\u0000fails.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Torchtree: flexible phylogenetic model development and inference using PyTorch Torchtree:使用 PyTorch 进行灵活的系统发生模型开发和推断
Pub Date : 2024-06-26 DOI: arxiv-2406.18044
Mathieu Fourment, Matthew Macaulay, Christiaan J Swanepoel, Xiang Ji, Marc A Suchard, Frederick A Matsen IV
Bayesian inference has predominantly relied on the Markov chain Monte Carlo(MCMC) algorithm for many years. However, MCMC is computationally laborious,especially for complex phylogenetic models of time trees. This bottleneck hasled to the search for alternatives, such as variational Bayes, which can scalebetter to large datasets. In this paper, we introduce torchtree, a frameworkwritten in Python that allows developers to easily implement rich phylogeneticmodels and algorithms using a fixed tree topology. One can either use automaticdifferentiation, or leverage torchtree's plug-in system to compute gradientsanalytically for model components for which automatic differentiation is slow.We demonstrate that the torchtree variational inference framework performssimilarly to BEAST in terms of speed and approximation accuracy. Furthermore,we explore the use of the forward KL divergence as an optimizing criterion forvariational inference, which can handle discontinuous and non-differentiablemodels. Our experiments show that inference using the forward KL divergencetends to be faster per iteration compared to the evidence lower bound (ELBO)criterion, although the ELBO-based inference may converge faster in some cases.Overall, torchtree provides a flexible and efficient framework for phylogeneticmodel development and inference using PyTorch.
多年来,贝叶斯推断主要依赖于马尔科夫链蒙特卡罗(MCMC)算法。然而,MCMC 计算起来非常费力,尤其是对于复杂的时间树系统发育模型。这一瓶颈导致人们开始寻找能更好地扩展到大型数据集的替代算法,如变异贝叶斯算法。在本文中,我们介绍了 torchtree,这是一个用 Python 编写的框架,允许开发人员使用固定的树拓扑结构轻松实现丰富的系统发育模型和算法。我们证明了 torchtree 变分推理框架在速度和近似精度方面的表现与 BEAST 相似。此外,我们还探索了使用前向 KL 发散作为变量推理的优化准则,它可以处理不连续和不可微分模型。我们的实验表明,与证据下限(ELBO)准则相比,使用前向 KL 发散进行推理的每次迭代速度更快,尽管基于 ELBO 的推理在某些情况下收敛得更快。
{"title":"Torchtree: flexible phylogenetic model development and inference using PyTorch","authors":"Mathieu Fourment, Matthew Macaulay, Christiaan J Swanepoel, Xiang Ji, Marc A Suchard, Frederick A Matsen IV","doi":"arxiv-2406.18044","DOIUrl":"https://doi.org/arxiv-2406.18044","url":null,"abstract":"Bayesian inference has predominantly relied on the Markov chain Monte Carlo\u0000(MCMC) algorithm for many years. However, MCMC is computationally laborious,\u0000especially for complex phylogenetic models of time trees. This bottleneck has\u0000led to the search for alternatives, such as variational Bayes, which can scale\u0000better to large datasets. In this paper, we introduce torchtree, a framework\u0000written in Python that allows developers to easily implement rich phylogenetic\u0000models and algorithms using a fixed tree topology. One can either use automatic\u0000differentiation, or leverage torchtree's plug-in system to compute gradients\u0000analytically for model components for which automatic differentiation is slow.\u0000We demonstrate that the torchtree variational inference framework performs\u0000similarly to BEAST in terms of speed and approximation accuracy. Furthermore,\u0000we explore the use of the forward KL divergence as an optimizing criterion for\u0000variational inference, which can handle discontinuous and non-differentiable\u0000models. Our experiments show that inference using the forward KL divergence\u0000tends to be faster per iteration compared to the evidence lower bound (ELBO)\u0000criterion, although the ELBO-based inference may converge faster in some cases.\u0000Overall, torchtree provides a flexible and efficient framework for phylogenetic\u0000model development and inference using PyTorch.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scalable Sampling of Truncated Multivariate Normals Using Sequential Nearest-Neighbor Approximation 利用序列近邻逼近对截断多变量正态进行可扩展采样
Pub Date : 2024-06-25 DOI: arxiv-2406.17307
Jian Cao, Matthias Katzfuss
We propose a linear-complexity method for sampling from truncatedmultivariate normal (TMVN) distributions with high fidelity by applyingnearest-neighbor approximations to a product-of-conditionals decomposition ofthe TMVN density. To make the sequential sampling based on the decompositionfeasible, we introduce a novel method that avoids the intractablehigh-dimensional TMVN distribution by sampling sequentially from$m$-dimensional TMVN distributions, where $m$ is a tuning parameter controllingthe fidelity. This allows us to overcome the existing methods' crucial problemof rapidly decreasing acceptance rates for increasing dimension. Throughout ourexperiments with up to tens of thousands of dimensions, we can producehigh-fidelity samples with $m$ in the dozens, achieving superior scalabilitycompared to existing state-of-the-art methods. We study a tetrachloroethyleneconcentration dataset that has $3{,}971$ observed responses and $20{,}730$undetected responses, together modeled as a partially censored Gaussianprocess, where our method enables posterior inference for the censoredresponses through sampling a $20{,}730$-dimensional TMVN distribution.
我们提出了一种线性复杂度方法,通过对截断多变量正态分布(TMVN)密度的条件乘积分解应用最近邻近似,从截断多变量正态分布中进行高保真采样。为了使基于分解的顺序采样可行,我们引入了一种新方法,通过从 $m$ 维 TMVN 分布(其中 $m$ 是控制保真度的调整参数)顺序采样,避免了难以处理的高维 TMVN 分布。这使我们克服了现有方法的关键问题,即随着维度的增加,接受率迅速降低。在我们进行的多达数万维度的实验中,我们可以生成 $m$ 为几十的高保真样本,与现有的最先进方法相比,我们实现了卓越的可扩展性。我们研究了一个四氯乙烯浓度数据集,该数据集有3{,}971$观测到的响应和20{,}730$未检测到的响应,这些响应一起被建模为部分删减的高斯过程(partially censored Gaussianprocess),我们的方法通过对20{,}730$维的TMVN分布进行采样,实现了对删减响应的后验推断。
{"title":"Scalable Sampling of Truncated Multivariate Normals Using Sequential Nearest-Neighbor Approximation","authors":"Jian Cao, Matthias Katzfuss","doi":"arxiv-2406.17307","DOIUrl":"https://doi.org/arxiv-2406.17307","url":null,"abstract":"We propose a linear-complexity method for sampling from truncated\u0000multivariate normal (TMVN) distributions with high fidelity by applying\u0000nearest-neighbor approximations to a product-of-conditionals decomposition of\u0000the TMVN density. To make the sequential sampling based on the decomposition\u0000feasible, we introduce a novel method that avoids the intractable\u0000high-dimensional TMVN distribution by sampling sequentially from\u0000$m$-dimensional TMVN distributions, where $m$ is a tuning parameter controlling\u0000the fidelity. This allows us to overcome the existing methods' crucial problem\u0000of rapidly decreasing acceptance rates for increasing dimension. Throughout our\u0000experiments with up to tens of thousands of dimensions, we can produce\u0000high-fidelity samples with $m$ in the dozens, achieving superior scalability\u0000compared to existing state-of-the-art methods. We study a tetrachloroethylene\u0000concentration dataset that has $3{,}971$ observed responses and $20{,}730$\u0000undetected responses, together modeled as a partially censored Gaussian\u0000process, where our method enables posterior inference for the censored\u0000responses through sampling a $20{,}730$-dimensional TMVN distribution.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141523218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Genealogical processes of non-neutral population models under rapid mutation 快速突变下非中性种群模型的谱系过程
Pub Date : 2024-06-24 DOI: arxiv-2406.16465
Jere Koskela, Paul A. Jenkins, Adam M. Johansen, Dario Spano
We show that genealogical trees arising from a broad class of non-neutralmodels of population evolution converge to the Kingman coalescent under asuitable rescaling of time. As well as non-neutral biological evolution, ourresults apply to genetic algorithms encompassing the prominent class ofsequential Monte Carlo (SMC) methods. The time rescaling we need differsslightly from that used in classical results for convergence to the Kingmancoalescent, which has implications for the performance of different resamplingschemes in SMC algorithms. In addition, our work substantially simplifiesearlier proofs of convergence to the Kingman coalescent, and corrects an errorcommon to several earlier results.
我们的研究表明,在适当的时间重定标条件下,由一大类非中性种群进化模型产生的系谱树会向金曼聚合收敛。除了非中性生物进化,我们的结果还适用于遗传算法,包括著名的连续蒙特卡罗(SMC)方法。我们所需的时间重定标与经典的金曼科尺度收敛结果所使用的时间重定标略有不同,这对 SMC 算法中不同重采样策略的性能有影响。此外,我们的工作还大大简化了早先关于收敛到 Kingmancoalescent 的证明,并纠正了早先几个结果中常见的错误。
{"title":"Genealogical processes of non-neutral population models under rapid mutation","authors":"Jere Koskela, Paul A. Jenkins, Adam M. Johansen, Dario Spano","doi":"arxiv-2406.16465","DOIUrl":"https://doi.org/arxiv-2406.16465","url":null,"abstract":"We show that genealogical trees arising from a broad class of non-neutral\u0000models of population evolution converge to the Kingman coalescent under a\u0000suitable rescaling of time. As well as non-neutral biological evolution, our\u0000results apply to genetic algorithms encompassing the prominent class of\u0000sequential Monte Carlo (SMC) methods. The time rescaling we need differs\u0000slightly from that used in classical results for convergence to the Kingman\u0000coalescent, which has implications for the performance of different resampling\u0000schemes in SMC algorithms. In addition, our work substantially simplifies\u0000earlier proofs of convergence to the Kingman coalescent, and corrects an error\u0000common to several earlier results.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141523192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Recursive variational Gaussian approximation with the Whittle likelihood for linear non-Gaussian state space models 利用惠特尔似然对线性非高斯状态空间模型进行递归变分高斯逼近
Pub Date : 2024-06-23 DOI: arxiv-2406.15998
Bao Anh Vu, David Gunawan, Andrew Zammit-Mangion
Parameter inference for linear and non-Gaussian state space models ischallenging because the likelihood function contains an intractable integralover the latent state variables. Exact inference using Markov chain Monte Carlois computationally expensive, particularly for long time series data.Variational Bayes methods are useful when exact inference is infeasible. Thesemethods approximate the posterior density of the parameters by a simple andtractable distribution found through optimisation. In this paper, we propose anovel sequential variational Bayes approach that makes use of the Whittlelikelihood for computationally efficient parameter inference in this class ofstate space models. Our algorithm, which we call Recursive Variational GaussianApproximation with the Whittle Likelihood (R-VGA-Whittle), updates thevariational parameters by processing data in the frequency domain. At eachiteration, R-VGA-Whittle requires the gradient and Hessian of the Whittlelog-likelihood, which are available in closed form for a wide class of models.Through several examples using a linear Gaussian state space model and aunivariate/bivariate non-Gaussian stochastic volatility model, we show thatR-VGA-Whittle provides good approximations to posterior distributions of theparameters and is very computationally efficient when compared toasymptotically exact methods such as Hamiltonian Monte Carlo.
线性和非高斯状态空间模型的参数推断是一项挑战,因为似然函数包含一个难以处理的潜在状态变量积分。使用马尔科夫链蒙特卡罗进行精确推断的计算成本很高,尤其是对于长时间序列数据。当精确推断不可行时,变分贝叶斯方法就会派上用场。这些方法通过优化找到一个简单、可操作的分布,从而近似得到参数的后验密度。在本文中,我们提出了一种新的序列变分贝叶斯方法,该方法利用惠特尔似然(Whittlelikelihood)对这类状态空间模型中的参数进行高效计算推断。我们将这种算法称为 "惠特尔似然递归变异高斯逼近算法"(R-VGA-Whittle),它通过处理频域数据来更新变异参数。通过使用线性高斯状态空间模型和单变量/双变量非高斯随机波动性模型的几个例子,我们表明 R-VGA-Whittle 可以很好地近似参数的后验分布,与汉密尔顿蒙特卡洛等渐近精确方法相比,计算效率非常高。
{"title":"Recursive variational Gaussian approximation with the Whittle likelihood for linear non-Gaussian state space models","authors":"Bao Anh Vu, David Gunawan, Andrew Zammit-Mangion","doi":"arxiv-2406.15998","DOIUrl":"https://doi.org/arxiv-2406.15998","url":null,"abstract":"Parameter inference for linear and non-Gaussian state space models is\u0000challenging because the likelihood function contains an intractable integral\u0000over the latent state variables. Exact inference using Markov chain Monte Carlo\u0000is computationally expensive, particularly for long time series data.\u0000Variational Bayes methods are useful when exact inference is infeasible. These\u0000methods approximate the posterior density of the parameters by a simple and\u0000tractable distribution found through optimisation. In this paper, we propose a\u0000novel sequential variational Bayes approach that makes use of the Whittle\u0000likelihood for computationally efficient parameter inference in this class of\u0000state space models. Our algorithm, which we call Recursive Variational Gaussian\u0000Approximation with the Whittle Likelihood (R-VGA-Whittle), updates the\u0000variational parameters by processing data in the frequency domain. At each\u0000iteration, R-VGA-Whittle requires the gradient and Hessian of the Whittle\u0000log-likelihood, which are available in closed form for a wide class of models.\u0000Through several examples using a linear Gaussian state space model and a\u0000univariate/bivariate non-Gaussian stochastic volatility model, we show that\u0000R-VGA-Whittle provides good approximations to posterior distributions of the\u0000parameters and is very computationally efficient when compared to\u0000asymptotically exact methods such as Hamiltonian Monte Carlo.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - STAT - Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1