Convergence properties of incremental Bayesian evolutionary algorithms with single Markov chains

Byoung-Tak Zhang, G. Paass, H. Mühlenbein
{"title":"Convergence properties of incremental Bayesian evolutionary algorithms with single Markov chains","authors":"Byoung-Tak Zhang, G. Paass, H. Mühlenbein","doi":"10.1109/CEC.2000.870744","DOIUrl":null,"url":null,"abstract":"Bayesian evolutionary algorithms (BEAs) are a probabilistic model of evolutionary computation for learning and optimization. Starting from a population of individuals drawn from a prior distribution, a Bayesian evolutionary algorithm iteratively generates a new population by estimating the posterior fitness distribution of parent individuals and then sampling from the distribution offspring individuals by variation and selection operators. Due to the non-homogeneity of their Markov chains, the convergence properties of the full BEAs are difficult to analyze. However, recent developments in Markov chain analysis for dynamic Monte Carlo methods provide a useful tool for studying asymptotic behaviors of adaptive Markov chain Monte Carlo methods including evolutionary algorithms. We apply these results to Investigate the convergence properties of Bayesian evolutionary algorithms with incremental data growth. We study the case of BEAs that generate single chains or have populations of size one. It is shown that under regularity conditions the incremental BEA asymptotically converges to a maximum a posteriori (MAP) estimate which is concentrated around the maximum likelihood estimate. This result relies on the observation that increasing the number of data items has an equivalent effect of reducing the temperature in simulated annealing.","PeriodicalId":218136,"journal":{"name":"Proceedings of the 2000 Congress on Evolutionary Computation. CEC00 (Cat. No.00TH8512)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2000-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2000 Congress on Evolutionary Computation. CEC00 (Cat. No.00TH8512)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CEC.2000.870744","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

Abstract

Bayesian evolutionary algorithms (BEAs) are a probabilistic model of evolutionary computation for learning and optimization. Starting from a population of individuals drawn from a prior distribution, a Bayesian evolutionary algorithm iteratively generates a new population by estimating the posterior fitness distribution of parent individuals and then sampling from the distribution offspring individuals by variation and selection operators. Due to the non-homogeneity of their Markov chains, the convergence properties of the full BEAs are difficult to analyze. However, recent developments in Markov chain analysis for dynamic Monte Carlo methods provide a useful tool for studying asymptotic behaviors of adaptive Markov chain Monte Carlo methods including evolutionary algorithms. We apply these results to Investigate the convergence properties of Bayesian evolutionary algorithms with incremental data growth. We study the case of BEAs that generate single chains or have populations of size one. It is shown that under regularity conditions the incremental BEA asymptotically converges to a maximum a posteriori (MAP) estimate which is concentrated around the maximum likelihood estimate. This result relies on the observation that increasing the number of data items has an equivalent effect of reducing the temperature in simulated annealing.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
单马尔可夫链增量贝叶斯进化算法的收敛性
贝叶斯进化算法(BEAs)是一种用于学习和优化的概率进化计算模型。贝叶斯进化算法从一个先验分布中得到一个个体群体,通过估计亲本个体的后验适应度分布,迭代生成一个新的群体,然后通过变异算子和选择算子从分布的后代个体中抽样。由于其马尔可夫链的非齐次性,使得完整bea的收敛性难以分析。然而,动态蒙特卡罗方法中马尔可夫链分析的最新进展为研究包括进化算法在内的自适应马尔可夫链蒙特卡罗方法的渐近行为提供了有用的工具。我们应用这些结果来研究数据增量增长下贝叶斯进化算法的收敛性。我们研究了产生单链或具有大小为1的种群的BEAs的情况。结果表明,在正则性条件下,增量BEA渐近收敛于一个以最大似然估计为中心的最大后验估计。这一结果依赖于在模拟退火中增加数据项的数量具有降低温度的等效效果的观察。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Test-case generator TCG-2 for nonlinear parameter optimisation Accelerating multi-objective control system design using a neuro-genetic approach On the use of stochastic estimator learning automata for dynamic channel allocation in broadcast networks A hierarchical distributed genetic algorithm for image segmentation Genetic learning of multi-attribute interactions in speaker verification
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1