首页 > 最新文献

arXiv: Computation最新文献

英文 中文
Double Happiness: Enhancing the Coupled Gains of L-lag Coupling via Control Variates. 双喜:通过控制变量增强l -滞后耦合的耦合增益。
Pub Date : 2020-08-28 DOI: 10.5705/SS.202020.0461
Radu V. Craiu, X. Meng
The recently proposed L-lag coupling for unbiased MCMC citep{biswas2019estimating, jacob2020unbiased} calls for a joint celebration by MCMC practitioners and theoreticians. For practitioners, it circumvents the thorny issue of deciding the burn-in period or when to terminate an MCMC iteration, and opens the door for safe parallel implementation. For theoreticians, it provides a powerful tool to establish elegant and easily estimable bounds on the exact error of MCMC approximation at any finite number of iteration. A serendipitous observation about the bias correcting term led us to introduce naturally available control variates into the L-lag coupling estimators. In turn, this extension enhances the coupled gains of L-lag coupling, because it results in more efficient unbiased estimators as well as a better bound on the total variation error of MCMC iterations, albeit the gains diminish with the numerical value of L. Specifically, the new bound is theoretically guaranteed to never exceed the one given previously. We also argue that L-lag coupling represents a long sought after coupling for the future, breaking a logjam of the coupling-from-the-past type of perfect sampling, by reducing the generally un-achievable requirement of being textit{perfect} to being textit{unbiased}, a worthwhile trade-off for ease of implementation in most practical situations. The theoretical analysis is supported by numerical experiments that show tighter bounds and a gain in efficiency when control variates are introduced.
最近提出的无偏MCMC citep{biswas2019estimating, jacob2020unbiased}的l滞后耦合需要MCMC实践者和理论家共同庆祝。对于实践者来说,它规避了决定老化期或何时终止MCMC迭代的棘手问题,并为安全的并行实现打开了大门。对于理论家来说,它提供了一个强大的工具,可以在任意有限次迭代下建立精确误差的优雅且易于估计的边界。对偏差校正项的偶然观察使我们将自然可用的控制变量引入到l滞后耦合估计器中。反过来,这种扩展增强了L-lag耦合的耦合增益,因为它产生了更有效的无偏估计器以及MCMC迭代的总变异误差的更好的界,尽管增益随着l的数值而减小。具体来说,理论上保证新的界永远不会超过先前给出的界。我们还认为,l滞后耦合代表了未来长期追求的耦合,通过将通常无法实现的要求从textit{完美}降低到textit{不偏不倚},打破了过去完美采样类型耦合的僵局,这是在大多数实际情况下易于实现的值得权衡的代价。数值实验结果支持了理论分析,表明在引入控制变量时边界更紧,效率更高。
{"title":"Double Happiness: Enhancing the Coupled Gains of L-lag Coupling via Control Variates.","authors":"Radu V. Craiu, X. Meng","doi":"10.5705/SS.202020.0461","DOIUrl":"https://doi.org/10.5705/SS.202020.0461","url":null,"abstract":"The recently proposed L-lag coupling for unbiased MCMC citep{biswas2019estimating, jacob2020unbiased} calls for a joint celebration by MCMC practitioners and theoreticians. For practitioners, it circumvents the thorny issue of deciding the burn-in period or when to terminate an MCMC iteration, and opens the door for safe parallel implementation. For theoreticians, it provides a powerful tool to establish elegant and easily estimable bounds on the exact error of MCMC approximation at any finite number of iteration. A serendipitous observation about the bias correcting term led us to introduce naturally available control variates into the L-lag coupling estimators. In turn, this extension enhances the coupled gains of L-lag coupling, because it results in more efficient unbiased estimators as well as a better bound on the total variation error of MCMC iterations, albeit the gains diminish with the numerical value of L. Specifically, the new bound is theoretically guaranteed to never exceed the one given previously. We also argue that L-lag coupling represents a long sought after coupling for the future, breaking a logjam of the coupling-from-the-past type of perfect sampling, by reducing the generally un-achievable requirement of being textit{perfect} to being textit{unbiased}, a worthwhile trade-off for ease of implementation in most practical situations. The theoretical analysis is supported by numerical experiments that show tighter bounds and a gain in efficiency when control variates are introduced.","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"72 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84090470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
SCOREDRIVENMODELS.JL: A JULIA PACKAGE FOR GENERALIZED AUTOREGRESSIVE SCORE MODELS SCOREDRIVENMODELS。一个用于广义自回归分数模型的Julia包
Pub Date : 2020-08-12 DOI: 10.17771/pucrio.acad.57291
Guilherme Bodin, Raphael Saavedra, C. Fernandes, A. Street
Score-driven models, also known as generalized autoregressive score models, represent a class of observation-driven time series models. They possess powerful properties, such as the ability to model different conditional distributions and to consider time-varying parameters within a flexible framework. In this paper, we present ScoreDrivenModels.jl, an open-source Julia package for modeling, forecasting, and simulating time series using the framework of score-driven models. The package is flexible with respect to model definition, allowing the user to specify the lag structure and which parameters are time-varying or constant. It is also possible to consider several distributions, including Beta, Exponential, Gamma, Lognormal, Normal, Poisson, Student's t, and Weibull. The provided interface is flexible, allowing interested users to implement any desired distribution and parametrization.
分数驱动模型,也称为广义自回归分数模型,代表了一类观测驱动的时间序列模型。它们具有强大的特性,例如能够对不同的条件分布进行建模,并在灵活的框架内考虑时变参数。在本文中,我们提出了ScoreDrivenModels。jl,一个开源的Julia包,用于使用分数驱动模型的框架来建模、预测和模拟时间序列。该包在模型定义方面是灵活的,允许用户指定滞后结构以及哪些参数随时间变化或恒定。也可以考虑几种分布,包括Beta、指数、Gamma、对数正态、正态、泊松、Student’st和威布尔。所提供的接口是灵活的,允许感兴趣的用户实现任何期望的分布和参数化。
{"title":"SCOREDRIVENMODELS.JL: A JULIA PACKAGE FOR GENERALIZED AUTOREGRESSIVE SCORE MODELS","authors":"Guilherme Bodin, Raphael Saavedra, C. Fernandes, A. Street","doi":"10.17771/pucrio.acad.57291","DOIUrl":"https://doi.org/10.17771/pucrio.acad.57291","url":null,"abstract":"Score-driven models, also known as generalized autoregressive score models, represent a class of observation-driven time series models. They possess powerful properties, such as the ability to model different conditional distributions and to consider time-varying parameters within a flexible framework. In this paper, we present ScoreDrivenModels.jl, an open-source Julia package for modeling, forecasting, and simulating time series using the framework of score-driven models. The package is flexible with respect to model definition, allowing the user to specify the lag structure and which parameters are time-varying or constant. It is also possible to consider several distributions, including Beta, Exponential, Gamma, Lognormal, Normal, Poisson, Student's t, and Weibull. The provided interface is flexible, allowing interested users to implement any desired distribution and parametrization.","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"23 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90496658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Simple conditions for convergence of sequential Monte Carlo genealogies with applications 序贯蒙特卡罗谱收敛的简单条件及其应用
Pub Date : 2020-06-30 DOI: 10.1214/20-ejp561
Suzie Brown, P. A. Jenkins, A. M. Johansen, Jere Koskela
Sequential Monte Carlo algorithms are popular methods for approximating integrals in problems such as non-linear filtering and smoothing. Their performance depends strongly on the properties of an induced genealogical process. We present simple conditions under which the limiting process, as the number of particles grows, is a time-rescaled Kingman coalescent. We establish these conditions for standard sequential Monte Carlo with a broad class of low-variance resampling schemes, as well as for conditional sequential Monte Carlo with multinomial resampling.
时序蒙特卡罗算法是在非线性滤波和平滑等问题中逼近积分的常用方法。它们的表现在很大程度上取决于诱导谱系过程的性质。我们给出了一些简单的条件,在这些条件下,随着粒子数量的增加,极限过程是一个时间尺度的金曼聚结。我们为具有广泛的低方差重采样方案的标准序列蒙特卡罗,以及具有多项重采样的条件序列蒙特卡罗,建立了这些条件。
{"title":"Simple conditions for convergence of sequential Monte Carlo genealogies with applications","authors":"Suzie Brown, P. A. Jenkins, A. M. Johansen, Jere Koskela","doi":"10.1214/20-ejp561","DOIUrl":"https://doi.org/10.1214/20-ejp561","url":null,"abstract":"Sequential Monte Carlo algorithms are popular methods for approximating integrals in problems such as non-linear filtering and smoothing. Their performance depends strongly on the properties of an induced genealogical process. We present simple conditions under which the limiting process, as the number of particles grows, is a time-rescaled Kingman coalescent. We establish these conditions for standard sequential Monte Carlo with a broad class of low-variance resampling schemes, as well as for conditional sequential Monte Carlo with multinomial resampling.","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"14 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73418886","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Increasing the efficiency of Sequential Monte Carlo samplers through the use of approximately optimal L-kernels 通过使用近似最优l核来提高顺序蒙特卡罗采样器的效率
Pub Date : 2020-04-24 DOI: 10.1016/j.ymssp.2021.108028
P. L. Green, Robert E. Moore, Ryan J Jackson, Jinglai Li, S. Maskell
{"title":"Increasing the efficiency of Sequential Monte Carlo samplers through the use of approximately optimal L-kernels","authors":"P. L. Green, Robert E. Moore, Ryan J Jackson, Jinglai Li, S. Maskell","doi":"10.1016/j.ymssp.2021.108028","DOIUrl":"https://doi.org/10.1016/j.ymssp.2021.108028","url":null,"abstract":"","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"37 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76690681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Particle Methods for Stochastic Differential Equation Mixed Effects Models 随机微分方程混合效应模型的粒子方法
Pub Date : 2019-07-25 DOI: 10.1214/20-ba1216
Imke Botha, R. Kohn, C. Drovandi
Parameter inference for stochastic differential equation mixed effects models (SDEMEMs) is a challenging problem. Analytical solutions for these models are rarely available, which means that the likelihood is also intractable. In this case, exact inference is possible using the pseudo-marginal method, where the intractable likelihood is replaced by its nonnegative unbiased estimate. A useful application of this idea is particle MCMC, which uses a particle filter estimate of the likelihood. While the exact posterior is targeted by these methods, a naive implementation for SDEMEMs can be highly inefficient. We develop three extensions to the naive approach which exploits specific aspects of SDEMEMs and other advances such as correlated pseudo-marginal methods. We compare these methods on real and simulated data from a tumour xenography study on mice.
随机微分方程混合效应模型(SDEMEMs)的参数推断是一个具有挑战性的问题。这些模型的分析解决方案很少可用,这意味着可能性也是难以处理的。在这种情况下,使用伪边际方法可以进行精确推断,其中难以处理的似然被其非负无偏估计所取代。这个想法的一个有用的应用是粒子MCMC,它使用粒子滤波估计可能性。虽然这些方法的目标是精确的后验,但对于SDEMEMs的幼稚实现可能非常低效。我们开发了朴素方法的三个扩展,利用了SDEMEMs的特定方面和其他进展,如相关的伪边际方法。我们比较这些方法的真实和模拟数据从肿瘤异种研究的小鼠。
{"title":"Particle Methods for Stochastic Differential Equation Mixed Effects Models","authors":"Imke Botha, R. Kohn, C. Drovandi","doi":"10.1214/20-ba1216","DOIUrl":"https://doi.org/10.1214/20-ba1216","url":null,"abstract":"Parameter inference for stochastic differential equation mixed effects models (SDEMEMs) is a challenging problem. Analytical solutions for these models are rarely available, which means that the likelihood is also intractable. In this case, exact inference is possible using the pseudo-marginal method, where the intractable likelihood is replaced by its nonnegative unbiased estimate. A useful application of this idea is particle MCMC, which uses a particle filter estimate of the likelihood. While the exact posterior is targeted by these methods, a naive implementation for SDEMEMs can be highly inefficient. We develop three extensions to the naive approach which exploits specific aspects of SDEMEMs and other advances such as correlated pseudo-marginal methods. We compare these methods on real and simulated data from a tumour xenography study on mice.","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"6 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78647733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Nested sampling on non-trivial geometries 非平凡几何上的嵌套抽样
Pub Date : 2019-05-22 DOI: 10.5281/zenodo.3653182
K. Javid
Metropolis nested sampling evolves a Markov chain from a current livepoint and accepts new points along the chain according to a version of the Metropolis acceptance ratio modified to satisfy the likelihood constraint, characteristic of nested sampling algorithms. The geometric nested sampling algorithm we present here is a based on the Metropolis method, but treats parameters as though they represent points on certain geometric objects, namely circles, tori and spheres. For parameters which represent points on a circle or torus, the trial distribution is `wrapped' around the domain of the posterior distribution such that samples cannot be rejected automatically when evaluating the Metropolis ratio due to being outside the sampling domain. Furthermore, this enhances the mobility of the sampler. For parameters which represent coordinates on the surface of a sphere, the algorithm transforms the parameters into a Cartesian coordinate system before sampling which again makes sure no samples are automatically rejected, and provides a physically intutive way of the sampling the parameter space. We apply the geometric nested sampler to two types of toy model which include circular, toroidal and spherical parameters. We find that the geometric nested sampler generally outperforms textsc{MultiNest} in both cases. %We also apply the algorithm to a gravitational wave detection model which includes circular and spherical parameters, and find that the geometric nested sampler and textsc{MultiNest} appear to perform equally well as one another. Our implementation of the algorithm can be found at url{this https URL}.
Metropolis嵌套抽样从当前livepoint进化出一条马尔可夫链,并根据修改后的Metropolis接受比版本接受链上的新点,以满足嵌套抽样算法的似然约束。我们在这里提出的几何嵌套采样算法是基于Metropolis方法的,但将参数视为某些几何对象(即圆、环面和球体)上的点。对于代表圆形或环面上点的参数,试验分布被“包裹”在后验分布的域周围,这样在评估Metropolis比率时,由于在采样域之外,样本不能被自动拒绝。此外,这提高了采样器的流动性。对于表示球面坐标的参数,该算法在采样前将其转换为直角坐标系,再次保证了采样不会被自动拒绝,并提供了一种物理上直观的参数空间采样方式。我们将几何嵌套采样器应用于包括圆形、环面和球面参数的两类玩具模型。我们发现,在这两种情况下,几何嵌套采样器通常优于textsc{多项}测试。 %We also apply the algorithm to a gravitational wave detection model which includes circular and spherical parameters, and find that the geometric nested sampler and textsc{MultiNest} appear to perform equally well as one another. Our implementation of the algorithm can be found at url{this https URL}.
{"title":"Nested sampling on non-trivial geometries","authors":"K. Javid","doi":"10.5281/zenodo.3653182","DOIUrl":"https://doi.org/10.5281/zenodo.3653182","url":null,"abstract":"Metropolis nested sampling evolves a Markov chain from a current livepoint and accepts new points along the chain according to a version of the Metropolis acceptance ratio modified to satisfy the likelihood constraint, characteristic of nested sampling algorithms. The geometric nested sampling algorithm we present here is a based on the Metropolis method, but treats parameters as though they represent points on certain geometric objects, namely circles, tori and spheres. For parameters which represent points on a circle or torus, the trial distribution is `wrapped' around the domain of the posterior distribution such that samples cannot be rejected automatically when evaluating the Metropolis ratio due to being outside the sampling domain. Furthermore, this enhances the mobility of the sampler. For parameters which represent coordinates on the surface of a sphere, the algorithm transforms the parameters into a Cartesian coordinate system before sampling which again makes sure no samples are automatically rejected, and provides a physically intutive way of the sampling the parameter space. \u0000We apply the geometric nested sampler to two types of toy model which include circular, toroidal and spherical parameters. We find that the geometric nested sampler generally outperforms textsc{MultiNest} in both cases. %We also apply the algorithm to a gravitational wave detection model which includes circular and spherical parameters, and find that the geometric nested sampler and textsc{MultiNest} appear to perform equally well as one another. Our implementation of the algorithm can be found at url{this https URL}.","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"235 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-05-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75725631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Pushing the Limit: A Hybrid Parallel Implementation of the Multi-resolution Approximation for Massive Data 突破极限:海量数据多分辨率近似的混合并行实现
Pub Date : 2019-04-30 DOI: 10.5065/nnt6-q689
Huang Huang, Lewis R. Blake, D. Hammerling
The multi-resolution approximation (MRA) of Gaussian processes was recently proposed to conduct likelihood-based inference for massive spatial data sets. An advantage of the methodology is that it can be parallelized. We implemented the MRA in C++ for both serial and parallel versions. In the parallel implementation, we use a hybrid parallelism that employs both distributed and shared memory computing for communications between and within nodes by using the Message Passing Interface (MPI) and OpenMP, respectively. The performance of the serial code is compared between the C++ and MATLAB implementations over a small data set on a personal laptop. The C++ parallel program is further carefully studied under different configurations by applications to data sets from around a tenth of a million to 47 million observations. We show the practicality of this implementation by demonstrating that we can get quick inference for massive real-world data sets. The serial and parallel C++ code can be found at this https URL.
近年来提出了高斯过程的多分辨率近似(MRA),用于对海量空间数据集进行基于似然的推理。这种方法的一个优点是它可以并行化。我们在c++中实现了串行和并行版本的MRA。在并行实现中,我们使用混合并行,分别使用消息传递接口(Message Passing Interface, MPI)和OpenMP,在节点之间和节点内部使用分布式和共享内存计算进行通信。在个人笔记本电脑上的一个小数据集上,比较了c++和MATLAB实现串行代码的性能。c++并行程序在不同的配置下被进一步仔细研究,应用程序的数据集从一百万的十分之一到四千七百万的观测值。我们通过演示我们可以对大量真实数据集进行快速推理来展示这种实现的实用性。串行和并行c++代码可以在这个https URL中找到。
{"title":"Pushing the Limit: A Hybrid Parallel Implementation of the Multi-resolution Approximation for Massive Data","authors":"Huang Huang, Lewis R. Blake, D. Hammerling","doi":"10.5065/nnt6-q689","DOIUrl":"https://doi.org/10.5065/nnt6-q689","url":null,"abstract":"The multi-resolution approximation (MRA) of Gaussian processes was recently proposed to conduct likelihood-based inference for massive spatial data sets. An advantage of the methodology is that it can be parallelized. We implemented the MRA in C++ for both serial and parallel versions. In the parallel implementation, we use a hybrid parallelism that employs both distributed and shared memory computing for communications between and within nodes by using the Message Passing Interface (MPI) and OpenMP, respectively. The performance of the serial code is compared between the C++ and MATLAB implementations over a small data set on a personal laptop. The C++ parallel program is further carefully studied under different configurations by applications to data sets from around a tenth of a million to 47 million observations. We show the practicality of this implementation by demonstrating that we can get quick inference for massive real-world data sets. The serial and parallel C++ code can be found at this https URL.","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"68 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85795636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
On the use of approximate Bayesian computation Markov chain Monte Carlo with inflated tolerance and post-correction 利用近似贝叶斯计算马尔可夫链蒙特卡罗膨胀容限和后校正
Pub Date : 2019-02-01 DOI: 10.1093/biomet/asz078
M. Vihola, Jordan Franks
Approximate Bayesian computation allows for inference of complicated probabilistic models with intractable likelihoods using model simulations. The Markov chain Monte Carlo implementation of approximate Bayesian computation is often sensitive to the tolerance parameter: low tolerance leads to poor mixing and large tolerance entails excess bias. We consider an approach using a relatively large tolerance for the Markov chain Monte Carlo sampler to ensure its sufficient mixing, and post-processing the output leading to estimators for a range of finer tolerances. We introduce an approximate confidence interval for the related post-corrected estimators, and propose an adaptive approximate Bayesian computation Markov chain Monte Carlo, which finds a `balanced' tolerance level automatically, based on acceptance rate optimisation. Our experiments show that post-processing based estimators can perform better than direct Markov chain targetting a fine tolerance, that our confidence intervals are reliable, and that our adaptive algorithm leads to reliable inference with little user specification.
近似贝叶斯计算允许使用模型模拟来推断具有难以处理的可能性的复杂概率模型。近似贝叶斯计算的马尔可夫链蒙特卡罗实现往往对容差参数敏感:容差小导致混合差,容差大导致偏差过大。我们考虑了一种使用相对较大公差的马尔可夫链蒙特卡罗采样器的方法,以确保其充分混合,并对输出进行后处理,从而得到一系列更细公差的估计器。我们为相关的后校正估计量引入了一个近似置信区间,并提出了一种自适应近似贝叶斯计算马尔可夫链蒙特卡罗算法,该算法基于接受率优化自动找到一个“平衡”的容忍水平。我们的实验表明,基于后处理的估计器可以比直接马尔可夫链表现得更好,我们的置信区间是可靠的,并且我们的自适应算法可以在很少的用户规范下产生可靠的推断。
{"title":"On the use of approximate Bayesian computation Markov chain Monte Carlo with inflated tolerance and post-correction","authors":"M. Vihola, Jordan Franks","doi":"10.1093/biomet/asz078","DOIUrl":"https://doi.org/10.1093/biomet/asz078","url":null,"abstract":"Approximate Bayesian computation allows for inference of complicated probabilistic models with intractable likelihoods using model simulations. The Markov chain Monte Carlo implementation of approximate Bayesian computation is often sensitive to the tolerance parameter: low tolerance leads to poor mixing and large tolerance entails excess bias. We consider an approach using a relatively large tolerance for the Markov chain Monte Carlo sampler to ensure its sufficient mixing, and post-processing the output leading to estimators for a range of finer tolerances. We introduce an approximate confidence interval for the related post-corrected estimators, and propose an adaptive approximate Bayesian computation Markov chain Monte Carlo, which finds a `balanced' tolerance level automatically, based on acceptance rate optimisation. Our experiments show that post-processing based estimators can perform better than direct Markov chain targetting a fine tolerance, that our confidence intervals are reliable, and that our adaptive algorithm leads to reliable inference with little user specification.","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"48 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85157685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Precision Annealing Monte Carlo Methods for Statistical Data Assimilation: Metropolis-Hastings Procedures 统计数据同化的精密退火蒙特卡罗方法:大都会-黑斯廷斯程序
Pub Date : 2019-01-14 DOI: 10.5194/NPG-2019-1
Adrian S. Wong, Kangbo Hao, Zheng Fang, H. Abarbanel
Abstract. Statistical Data Assimilation (SDA) is the transfer of information from field or laboratory observations to a user selected model of the dynamical system producing those observations. The data is noisy and the model has errors; the information transfer addresses properties of the conditional probability distribution of the states of the model conditioned on the observations. The quantities of interest in SDA are the conditional expected values of functions of the model state, and these require the approximate evaluation of high dimensional integrals. We introduce a conditional probability distribution and use the Laplace method with annealing to identify the maxima of the conditional probability distribution. The annealing method slowly increases the precision term of the model as it enters the Laplace method. In this paper, we extend the idea of precision annealing (PA) to Monte Carlo calculations of conditional expected values using Metropolis-Hastings methods.
摘要统计数据同化(SDA)是将现场或实验室观测的信息传递到产生这些观测的用户选择的动力系统模型。数据有噪声,模型存在误差;信息传递处理了模型状态的条件概率分布的属性,这些属性以观测值为条件。SDA中感兴趣的量是模型状态函数的条件期望值,这些需要对高维积分进行近似评估。我们引入了一个条件概率分布,并使用拉普拉斯退火法来识别条件概率分布的最大值。退火方法在进入拉普拉斯方法时,缓慢地增加模型的精度项。本文利用Metropolis-Hastings方法,将精密退火(PA)的思想推广到条件期望值的蒙特卡罗计算中。
{"title":"Precision Annealing Monte Carlo Methods for Statistical Data Assimilation: Metropolis-Hastings Procedures","authors":"Adrian S. Wong, Kangbo Hao, Zheng Fang, H. Abarbanel","doi":"10.5194/NPG-2019-1","DOIUrl":"https://doi.org/10.5194/NPG-2019-1","url":null,"abstract":"Abstract. Statistical Data Assimilation (SDA) is the transfer of information from field or laboratory observations to a user selected model of the dynamical system producing those observations. The data is noisy and the model has errors; the information transfer addresses properties of the conditional probability distribution of the states of the model conditioned on the observations. The quantities of interest in SDA are the conditional expected values of functions of the model state, and these require the approximate evaluation of high dimensional integrals. We introduce a conditional probability distribution and use the Laplace method with annealing to identify the maxima of the conditional probability distribution. The annealing method slowly increases the precision term of the model as it enters the Laplace method. In this paper, we extend the idea of precision annealing (PA) to Monte Carlo calculations of conditional expected values using Metropolis-Hastings methods.\u0000","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"37 48 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82758601","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Sparse regression with Multi-type Regularized Feature modeling 基于多类型正则化特征建模的稀疏回归
Pub Date : 2018-10-07 DOI: 10.1016/j.insmatheco.2020.11.010
Sander Devriendt, Katrien Antonio, Tom Reynkens, Roel Verbelen
{"title":"Sparse regression with Multi-type Regularized Feature modeling","authors":"Sander Devriendt, Katrien Antonio, Tom Reynkens, Roel Verbelen","doi":"10.1016/j.insmatheco.2020.11.010","DOIUrl":"https://doi.org/10.1016/j.insmatheco.2020.11.010","url":null,"abstract":"","PeriodicalId":8446,"journal":{"name":"arXiv: Computation","volume":"76 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90512921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
期刊
arXiv: Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1