{"title":"Efficient nonparametric importance sampling for Bayesian learning","authors":"M. Zlochin, Y. Baram","doi":"10.1109/IJCNN.2002.1007535","DOIUrl":null,"url":null,"abstract":"Monte Carlo methods, such as importance sampling, have become a major tool in Bayesian inference. However, in order to produce an accurate estimate, the sampling distribution is required to be close to the target distribution. Several adaptive importance sampling algorithms, proposed over the last few years, attempt to learn a good sampling distribution automatically, but their performance is often unsatisfactory. In addition, a theoretical analysis, which takes into account the computational cost of the sampling algorithms, is still lacking. In this paper, we present a first attempt at such analysis, and we propose some modifications to existing adaptive importance sampling algorithms, which produce significantly more accurate estimates.","PeriodicalId":382771,"journal":{"name":"Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN'02 (Cat. No.02CH37290)","volume":"154 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2002-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN'02 (Cat. No.02CH37290)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN.2002.1007535","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
Monte Carlo methods, such as importance sampling, have become a major tool in Bayesian inference. However, in order to produce an accurate estimate, the sampling distribution is required to be close to the target distribution. Several adaptive importance sampling algorithms, proposed over the last few years, attempt to learn a good sampling distribution automatically, but their performance is often unsatisfactory. In addition, a theoretical analysis, which takes into account the computational cost of the sampling algorithms, is still lacking. In this paper, we present a first attempt at such analysis, and we propose some modifications to existing adaptive importance sampling algorithms, which produce significantly more accurate estimates.