首页 > 最新文献

J. Mach. Learn. Res.最新文献

英文 中文
Shark 鲨鱼
Pub Date : 2020-04-07 DOI: 10.5555/1390681.1390714
Christian Igel, V. Heidrich-Meisner, T. Glasmachers
{"title":"Shark","authors":"Christian Igel, V. Heidrich-Meisner, T. Glasmachers","doi":"10.5555/1390681.1390714","DOIUrl":"https://doi.org/10.5555/1390681.1390714","url":null,"abstract":"","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"45 1","pages":"993-996"},"PeriodicalIF":0.0,"publicationDate":"2020-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88380711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 125
Rankboost+: an improvement to Rankboost Rankboost+: Rankboost的改进
Pub Date : 2020-01-01 DOI: 10.1007/S10994-019-05826-X
H. Connamacher, Nikil Pancha, Rui Liu, Soumya Ray
{"title":"Rankboost+: an improvement to Rankboost","authors":"H. Connamacher, Nikil Pancha, Rui Liu, Soumya Ray","doi":"10.1007/S10994-019-05826-X","DOIUrl":"https://doi.org/10.1007/S10994-019-05826-X","url":null,"abstract":"","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"8 1","pages":"51-78"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78234568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Achieving Fairness in the Stochastic Multi-armed Bandit Problem 随机多臂盗匪问题公平性的实现
Pub Date : 2019-07-23 DOI: 10.1609/AAAI.V34I04.5986
Vishakha Patil, Ganesh Ghalme, V. Nair, Y. Narahari
We study an interesting variant of the stochastic multi-armed bandit problem, called the Fair-SMAB problem, where each arm is required to be pulled for at least a given fraction of the total available rounds. We investigate the interplay between learning and fairness in terms of a pre-specified vector denoting the fractions of guaranteed pulls. We define a fairness-aware regret, called $r$-Regret, that takes into account the above fairness constraints and naturally extends the conventional notion of regret. Our primary contribution is characterizing a class of Fair-SMAB algorithms by two parameters: the unfairness tolerance and the learning algorithm used as a black-box. We provide a fairness guarantee for this class that holds uniformly over time irrespective of the choice of the learning algorithm. In particular, when the learning algorithm is UCB1, we show that our algorithm achieves $O(ln T)$ $r$-Regret. Finally, we evaluate the cost of fairness in terms of the conventional notion of regret.
我们研究了随机多臂强盗问题的一个有趣的变体,称为Fair-SMAB问题,其中每只手臂被要求至少在总可用回合的给定分数内被拉动。我们研究了学习和公平之间的相互作用,根据一个预先指定的向量表示保证牵引力的分数。我们定义了一种公平意识的后悔,称为$r$-后悔,它考虑了上述公平约束,自然地扩展了传统的后悔概念。我们的主要贡献是通过两个参数来表征一类Fair-SMAB算法:不公平容忍度和用作黑盒的学习算法。我们为这个类提供了一个公平的保证,无论学习算法的选择如何,它都会随着时间的推移而保持一致。特别是,当学习算法为UCB1时,我们证明了我们的算法达到了$O(ln T)$ $r$-悔恨。最后,我们根据传统的后悔概念来评估公平的成本。
{"title":"Achieving Fairness in the Stochastic Multi-armed Bandit Problem","authors":"Vishakha Patil, Ganesh Ghalme, V. Nair, Y. Narahari","doi":"10.1609/AAAI.V34I04.5986","DOIUrl":"https://doi.org/10.1609/AAAI.V34I04.5986","url":null,"abstract":"We study an interesting variant of the stochastic multi-armed bandit problem, called the Fair-SMAB problem, where each arm is required to be pulled for at least a given fraction of the total available rounds. We investigate the interplay between learning and fairness in terms of a pre-specified vector denoting the fractions of guaranteed pulls. We define a fairness-aware regret, called $r$-Regret, that takes into account the above fairness constraints and naturally extends the conventional notion of regret. Our primary contribution is characterizing a class of Fair-SMAB algorithms by two parameters: the unfairness tolerance and the learning algorithm used as a black-box. We provide a fairness guarantee for this class that holds uniformly over time irrespective of the choice of the learning algorithm. In particular, when the learning algorithm is UCB1, we show that our algorithm achieves $O(ln T)$ $r$-Regret. Finally, we evaluate the cost of fairness in terms of the conventional notion of regret.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"15 1","pages":"174:1-174:31"},"PeriodicalIF":0.0,"publicationDate":"2019-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75610082","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 84
Multi-Player Bandits: The Adversarial Case 多人盗贼:对抗性案例
Pub Date : 2019-02-21 DOI: 10.3929/ETHZ-B-000414972
Pragnya Alatur, K. Levy, Andreas Krause
We consider a setting where multiple players sequentially choose among a common set of actions (arms). Motivated by a cognitive radio networks application, we assume that players incur a loss upon colliding, and that communication between players is not possible. Existing approaches assume that the system is stationary. Yet this assumption is often violated in practice, e.g., due to signal strength fluctuations. In this work, we design the first Multi-player Bandit algorithm that provably works in arbitrarily changing environments, where the losses of the arms may even be chosen by an adversary. This resolves an open problem posed by Rosenski, Shamir, and Szlak (2016).
我们考虑这样一个场景,即多个玩家依次选择一组共同的行动(武器)。受认知无线电网络应用程序的驱动,我们假设玩家在碰撞时遭受损失,并且玩家之间不可能进行交流。现有的方法假设系统是静止的。然而,这一假设在实践中经常被违反,例如,由于信号强度波动。在这项工作中,我们设计了第一个Multi-player Bandit算法,该算法可以在任意变化的环境中工作,其中手臂的损失甚至可以由对手选择。这解决了Rosenski、Shamir和Szlak(2016)提出的一个开放性问题。
{"title":"Multi-Player Bandits: The Adversarial Case","authors":"Pragnya Alatur, K. Levy, Andreas Krause","doi":"10.3929/ETHZ-B-000414972","DOIUrl":"https://doi.org/10.3929/ETHZ-B-000414972","url":null,"abstract":"We consider a setting where multiple players sequentially choose among a common set of actions (arms). Motivated by a cognitive radio networks application, we assume that players incur a loss upon colliding, and that communication between players is not possible. Existing approaches assume that the system is stationary. Yet this assumption is often violated in practice, e.g., due to signal strength fluctuations. In this work, we design the first Multi-player Bandit algorithm that provably works in arbitrarily changing environments, where the losses of the arms may even be chosen by an adversary. This resolves an open problem posed by Rosenski, Shamir, and Szlak (2016).","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"515 1","pages":"77:1-77:23"},"PeriodicalIF":0.0,"publicationDate":"2019-02-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77094807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
How Well Generative Adversarial Networks Learn Distributions 生成对抗网络如何学习分布
Pub Date : 2018-11-07 DOI: 10.2139/ssrn.3714011
Tengyuan Liang
This paper studies the rates of convergence for learning distributions implicitly with the adversarial framework and Generative Adversarial Networks (GAN), which subsume Wasserstein, Sobolev, MMD GAN, and Generalized/Simulated Method of Moments (GMM/SMM) as special cases. We study a wide range of parametric and nonparametric target distributions, under a host of objective evaluation metrics. We investigate how to obtain a good statistical guarantee for GANs through the lens of regularization. On the nonparametric end, we derive the optimal minimax rates for distribution estimation under the adversarial framework. On the parametric end, we establish a theory for general neural network classes (including deep leaky ReLU networks), that characterizes the interplay on the choice of generator and discriminator pair. We discover and isolate a new notion of regularization, called the generator-discriminator-pair regularization, that sheds light on the advantage of GANs compared to classical parametric and nonparametric approaches for explicit distribution estimation. We develop novel oracle inequalities as the main technical tools for analyzing GANs, which is of independent interest.
本文以Wasserstein、Sobolev、MMD GAN和广义/模拟矩法(GMM/SMM)为特例,研究了基于对抗框架和生成对抗网络的隐式学习分布的收敛速度。我们研究了一系列客观评价指标下的参数和非参数目标分布。我们从正则化的角度研究了如何获得gan的良好统计保证。在非参数端,我们导出了对抗性框架下分布估计的最优极大极小率。在参数端,我们建立了一般神经网络类(包括深度泄漏ReLU网络)的理论,表征了生成器和鉴别器对选择的相互作用。我们发现并分离了一个新的正则化概念,称为生成器-鉴别器对正则化,它揭示了gan与显式分布估计的经典参数和非参数方法相比的优势。我们开发了新的oracle不等式作为分析gan的主要技术工具,这是一个独立的兴趣。
{"title":"How Well Generative Adversarial Networks Learn Distributions","authors":"Tengyuan Liang","doi":"10.2139/ssrn.3714011","DOIUrl":"https://doi.org/10.2139/ssrn.3714011","url":null,"abstract":"This paper studies the rates of convergence for learning distributions implicitly with the adversarial framework and Generative Adversarial Networks (GAN), which subsume Wasserstein, Sobolev, MMD GAN, and Generalized/Simulated Method of Moments (GMM/SMM) as special cases. We study a wide range of parametric and nonparametric target distributions, under a host of objective evaluation metrics. We investigate how to obtain a good statistical guarantee for GANs through the lens of regularization. On the nonparametric end, we derive the optimal minimax rates for distribution estimation under the adversarial framework. On the parametric end, we establish a theory for general neural network classes (including deep leaky ReLU networks), that characterizes the interplay on the choice of generator and discriminator pair. We discover and isolate a new notion of regularization, called the generator-discriminator-pair regularization, that sheds light on the advantage of GANs compared to classical parametric and nonparametric approaches for explicit distribution estimation. We develop novel oracle inequalities as the main technical tools for analyzing GANs, which is of independent interest.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"103 1","pages":"228:1-228:41"},"PeriodicalIF":0.0,"publicationDate":"2018-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80303668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 64
Scikit-Multiflow: A Multi-output Streaming Framework Scikit-Multiflow:一个多输出流框架
Pub Date : 2018-07-01 DOI: 10.5555/3291125.3309634
Jacob Montiel, J. Read, A. Bifet, T. Abdessalem
Scikit-multiflow is a multi-output/multi-label and stream data mining framework for the Python programming language. Conceived to serve as a platform to encourage democratization of stream learning research, it provides multiple state of the art methods for stream learning, stream generators and evaluators. scikit-multiflow builds upon popular open source frameworks including scikit-learn, MOA and MEKA. Development follows the FOSS principles and quality is enforced by complying with PEP8 guidelines and using continuous integration and automatic testing. The source code is publicly available at this https URL.
Scikit-multiflow是Python编程语言的多输出/多标签和流数据挖掘框架。作为一个鼓励流学习研究民主化的平台,它为流学习、流生成器和评估器提供了多种最先进的方法。scikit-multiflow建立在流行的开源框架之上,包括scikit-learn、MOA和MEKA。开发遵循自由/开源软件原则,并通过遵守PEP8指导方针和使用持续集成和自动测试来强制执行质量。源代码可以在这个https URL上公开获得。
{"title":"Scikit-Multiflow: A Multi-output Streaming Framework","authors":"Jacob Montiel, J. Read, A. Bifet, T. Abdessalem","doi":"10.5555/3291125.3309634","DOIUrl":"https://doi.org/10.5555/3291125.3309634","url":null,"abstract":"Scikit-multiflow is a multi-output/multi-label and stream data mining framework for the Python programming language. Conceived to serve as a platform to encourage democratization of stream learning research, it provides multiple state of the art methods for stream learning, stream generators and evaluators. scikit-multiflow builds upon popular open source frameworks including scikit-learn, MOA and MEKA. Development follows the FOSS principles and quality is enforced by complying with PEP8 guidelines and using continuous integration and automatic testing. The source code is publicly available at this https URL.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"31 3","pages":"72:1-72:5"},"PeriodicalIF":0.0,"publicationDate":"2018-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91446714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 235
Deep Optimal Stopping 深度最佳停车
Pub Date : 2018-04-15 DOI: 10.3929/ETHZ-B-000344707
S. Becker, Patrick Cheridito, Arnulf Jentzen
In this paper we develop a deep learning method for optimal stopping problems which directly learns the optimal stopping rule from Monte Carlo samples. As such, it is broadly applicable in situations where the underlying randomness can efficiently be simulated. We test the approach on three problems: the pricing of a Bermudan max-call option, the pricing of a callable multi barrier reverse convertible and the problem of optimally stopping a fractional Brownian motion. In all three cases it produces very accurate results in high-dimensional situations with short computing times.
本文提出了一种直接从蒙特卡洛样本中学习最优停止规则的最优停止问题的深度学习方法。因此,它广泛适用于可以有效地模拟潜在随机性的情况。我们在三个问题上对该方法进行了测试:百慕大最大看涨期权的定价问题、可赎回多障碍反向可兑换期权的定价问题和最优停止分数布朗运动问题。在这三种情况下,它在高维情况下以较短的计算时间产生非常准确的结果。
{"title":"Deep Optimal Stopping","authors":"S. Becker, Patrick Cheridito, Arnulf Jentzen","doi":"10.3929/ETHZ-B-000344707","DOIUrl":"https://doi.org/10.3929/ETHZ-B-000344707","url":null,"abstract":"In this paper we develop a deep learning method for optimal stopping problems which directly learns the optimal stopping rule from Monte Carlo samples. As such, it is broadly applicable in situations where the underlying randomness can efficiently be simulated. We test the approach on three problems: the pricing of a Bermudan max-call option, the pricing of a callable multi barrier reverse convertible and the problem of optimally stopping a fractional Brownian motion. In all three cases it produces very accurate results in high-dimensional situations with short computing times.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"7 1","pages":"74:1-74:25"},"PeriodicalIF":0.0,"publicationDate":"2018-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87016819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 160
NeVAE: A Deep Generative Model for Molecular Graphs 分子图的深度生成模型
Pub Date : 2018-02-14 DOI: 10.1609/aaai.v33i01.33011110
Bidisha Samanta, A. De, G. Jana, P. Chattaraj, Niloy Ganguly, Manuel Gomez Rodriguez
Deep generative models have been praised for their ability to learn smooth latent representation of images, text, and audio, which can then be used to generate new, plausible data. However, current generative models are unable to work with molecular graphs due to their unique characteristics—their underlying structure is not Euclidean or grid-like, they remain isomorphic under permutation of the nodes labels, and they come with a different number of nodes and edges. In this paper, we propose NeVAE, a novel variational autoencoder for molecular graphs, whose encoder and decoder are specially designed to account for the above properties by means of several technical innovations. In addition, by using masking, the decoder is able to guarantee a set of valid properties in the generated molecules. Experiments reveal that our model can discover plausible, diverse and novel molecules more effectively than several state of the art methods. Moreover, by utilizing Bayesian optimization over the continuous latent representation of molecules our model finds, we can also find molecules that maximize certain desirable properties more effectively than alternatives.
深度生成模型因其学习图像、文本和音频的平滑潜在表示的能力而受到称赞,然后可用于生成新的、可信的数据。然而,目前的生成模型由于其独特的特性而无法处理分子图——它们的底层结构不是欧几里得或网格状的,它们在节点标签的排列下仍然是同构的,并且它们具有不同数量的节点和边。在本文中,我们提出了一种新的分子图变分自编码器neae,其编码器和解码器是通过一些技术创新而专门设计的。此外,通过使用掩蔽,解码器能够保证生成的分子中的一组有效属性。实验表明,我们的模型可以比几种最先进的方法更有效地发现合理的、多样的和新颖的分子。此外,通过对我们模型发现的分子的连续潜在表示使用贝叶斯优化,我们还可以找到比替代方案更有效地最大化某些理想特性的分子。
{"title":"NeVAE: A Deep Generative Model for Molecular Graphs","authors":"Bidisha Samanta, A. De, G. Jana, P. Chattaraj, Niloy Ganguly, Manuel Gomez Rodriguez","doi":"10.1609/aaai.v33i01.33011110","DOIUrl":"https://doi.org/10.1609/aaai.v33i01.33011110","url":null,"abstract":"Deep generative models have been praised for their ability to learn smooth latent representation of images, text, and audio, which can then be used to generate new, plausible data. However, current generative models are unable to work with molecular graphs due to their unique characteristics—their underlying structure is not Euclidean or grid-like, they remain isomorphic under permutation of the nodes labels, and they come with a different number of nodes and edges. In this paper, we propose NeVAE, a novel variational autoencoder for molecular graphs, whose encoder and decoder are specially designed to account for the above properties by means of several technical innovations. In addition, by using masking, the decoder is able to guarantee a set of valid properties in the generated molecules. Experiments reveal that our model can discover plausible, diverse and novel molecules more effectively than several state of the art methods. Moreover, by utilizing Bayesian optimization over the continuous latent representation of molecules our model finds, we can also find molecules that maximize certain desirable properties more effectively than alternatives.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"44 1","pages":"114:1-114:33"},"PeriodicalIF":0.0,"publicationDate":"2018-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88473595","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 177
Rate of Convergence of $k$-Nearest-Neighbor Classification Rule $k$-最近邻分类规则的收敛速度
Pub Date : 2017-10-16 DOI: 10.14760/OWP-2017-25
Maik Döring, L. Györfi, Harro Walk
A binary classification problem is considered. The excess error probability of the k-nearestneighbor classification rule according to the error probability of the Bayes decision is revisited by a decomposition of the excess error probability into approximation and estimation errors. Under a weak margin condition and under a modified Lipschitz condition or a local Lipschitz condition, tight upper bounds are presented such that one avoids the condition that the feature vector is bounded. The concept of modified Lipschitz condition is applied for discrete distributions, too. As a consequence of both concepts, we present the rate of convergence of L2 error for the corresponding nearest neighbor regression estimate.
考虑了一个二元分类问题。根据贝叶斯决策的错误概率,通过将超额错误概率分解为近似误差和估计误差,重新考察了k-最近邻分类规则的超额错误概率。在弱边界条件和改进的Lipschitz条件或局部Lipschitz条件下,给出了紧上界,从而避免了特征向量有界的条件。修正Lipschitz条件的概念也适用于离散分布。作为这两个概念的结果,我们给出了相应的最近邻回归估计的L2误差的收敛率。
{"title":"Rate of Convergence of $k$-Nearest-Neighbor Classification Rule","authors":"Maik Döring, L. Györfi, Harro Walk","doi":"10.14760/OWP-2017-25","DOIUrl":"https://doi.org/10.14760/OWP-2017-25","url":null,"abstract":"A binary classification problem is considered. The excess error probability of the k-nearestneighbor classification rule according to the error probability of the Bayes decision is revisited by a decomposition of the excess error probability into approximation and estimation errors. Under a weak margin condition and under a modified Lipschitz condition or a local Lipschitz condition, tight upper bounds are presented such that one avoids the condition that the feature vector is bounded. The concept of modified Lipschitz condition is applied for discrete distributions, too. As a consequence of both concepts, we present the rate of convergence of L2 error for the corresponding nearest neighbor regression estimate.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"27 1","pages":"227:1-227:16"},"PeriodicalIF":0.0,"publicationDate":"2017-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81261273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Covariances, Robustness, and Variational Bayes 协方差、稳健性和变分贝叶斯
Pub Date : 2017-09-08 DOI: 10.5555/3291125.3309613
Ryan Giordano, Tamara Broderick, Michael I. Jordan
Variational Bayes (VB) is an approximate Bayesian posterior inference technique that is increasingly popular due to its fast runtimes on large-scale datasets. However, even when VB provides accurate posterior means for certain parameters, it often mis-estimates variances and covariances. Furthermore, prior robustness measures have remained undeveloped for VB. By deriving a simple formula for the effect of infinitesimal model perturbations on VB posterior means, we provide both improved covariance estimates and local robustness measures for VB, thus greatly expanding the practical usefulness of VB posterior approximations. The estimates for VB posterior covariances rely on a result from the classical Bayesian robustness literature relating derivatives of posterior expectations to posterior covariances. Our key assumption is that the VB approximation provides good estimates of a select subset of posterior means -- an assumption that has been shown to hold in many practical settings. In our experiments, we demonstrate that our methods are simple, general, and fast, providing accurate posterior uncertainty estimates and robustness measures with runtimes that can be an order of magnitude smaller than MCMC.
变分贝叶斯(VB)是一种近似贝叶斯后验推理技术,由于其在大规模数据集上的快速运行而越来越受欢迎。然而,即使VB为某些参数提供了准确的后验均值,它也经常会错误地估计方差和协方差。此外,先前的健壮性措施还没有为VB开发。通过推导出无穷小模型扰动对VB后验均值影响的简单公式,我们为VB提供了改进的协方差估计和局部鲁棒性度量,从而极大地扩展了VB后验近似的实际用途。VB后验协方差的估计依赖于经典贝叶斯稳健性文献的结果,这些文献涉及后验期望对后验协方差的导数。我们的关键假设是,VB近似提供了后验均值选择子集的良好估计——这一假设已被证明在许多实际设置中成立。在我们的实验中,我们证明了我们的方法简单,通用,快速,提供准确的后验不确定性估计和鲁棒性测量,其运行时间可以比MCMC小一个数量级。
{"title":"Covariances, Robustness, and Variational Bayes","authors":"Ryan Giordano, Tamara Broderick, Michael I. Jordan","doi":"10.5555/3291125.3309613","DOIUrl":"https://doi.org/10.5555/3291125.3309613","url":null,"abstract":"Variational Bayes (VB) is an approximate Bayesian posterior inference technique that is increasingly popular due to its fast runtimes on large-scale datasets. However, even when VB provides accurate posterior means for certain parameters, it often mis-estimates variances and covariances. Furthermore, prior robustness measures have remained undeveloped for VB. By deriving a simple formula for the effect of infinitesimal model perturbations on VB posterior means, we provide both improved covariance estimates and local robustness measures for VB, thus greatly expanding the practical usefulness of VB posterior approximations. The estimates for VB posterior covariances rely on a result from the classical Bayesian robustness literature relating derivatives of posterior expectations to posterior covariances. Our key assumption is that the VB approximation provides good estimates of a select subset of posterior means -- an assumption that has been shown to hold in many practical settings. In our experiments, we demonstrate that our methods are simple, general, and fast, providing accurate posterior uncertainty estimates and robustness measures with runtimes that can be an order of magnitude smaller than MCMC.","PeriodicalId":14794,"journal":{"name":"J. Mach. Learn. Res.","volume":"70 1","pages":"51:1-51:49"},"PeriodicalIF":0.0,"publicationDate":"2017-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73755059","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 83
期刊
J. Mach. Learn. Res.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1