动态多目标优化问题解的增量细化

C. E. Mariano-Romero, M.E.F. Morales
{"title":"动态多目标优化问题解的增量细化","authors":"C. E. Mariano-Romero, M.E.F. Morales","doi":"10.1109/MICAI.2007.47","DOIUrl":null,"url":null,"abstract":"MDQL is an algorithm, based on reinforcement learning, for solving multiple objective optimization problems, that has been tested on several applications with promising results. MDQL discretizes the decision variables into a set of states, each associated with actions to move agents to contiguous states. A group of agents explore this state space and are able to find Pareto sets applying a distributed reinforcement learning algorithm. The precision of the Pareto solutions depends on the chosen granularity of the states. A finer granularity on the states creates more precise solutions but at the expense of a larger search space, and consequently the need for more computational resources. In this paper, a very important improvement is introduced into the original MDQL algorithm to incrementally refined the Pareto solutions. The new algorithm, called IMDQL, starts with a coarse granularity to find an initial Pareto set. A vicinity for each of the Pareto solutions in refined and a new Pareto set is founded in this refined state space. This process continues until there is no more improvement within a small threshold value. It is shown that IMDQL not only improves the solutions found by MDQL, but also converges faster. MDQL has also been tested on the solutions of dynamic optimization problems. In this paper, it is also shown that the adaptation capabilities observed in MDQL can be improved with IMDQL. IMDQL was tested on the benchmark problems proposed by Jin. Performance evaluation was made using the Collective Mean Fitness metric proposed by Morrison. IMDQL was compared against an standard evolution strategy with the covariance matrix adaptation (CMA-ES) with very promising results.","PeriodicalId":296192,"journal":{"name":"2007 Sixth Mexican International Conference on Artificial Intelligence, Special Session (MICAI)","volume":"88 4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Incremental Refinement of Solutions for Dynamic Multi Objective Optimization Problems\",\"authors\":\"C. E. Mariano-Romero, M.E.F. Morales\",\"doi\":\"10.1109/MICAI.2007.47\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"MDQL is an algorithm, based on reinforcement learning, for solving multiple objective optimization problems, that has been tested on several applications with promising results. MDQL discretizes the decision variables into a set of states, each associated with actions to move agents to contiguous states. A group of agents explore this state space and are able to find Pareto sets applying a distributed reinforcement learning algorithm. The precision of the Pareto solutions depends on the chosen granularity of the states. A finer granularity on the states creates more precise solutions but at the expense of a larger search space, and consequently the need for more computational resources. In this paper, a very important improvement is introduced into the original MDQL algorithm to incrementally refined the Pareto solutions. The new algorithm, called IMDQL, starts with a coarse granularity to find an initial Pareto set. A vicinity for each of the Pareto solutions in refined and a new Pareto set is founded in this refined state space. This process continues until there is no more improvement within a small threshold value. It is shown that IMDQL not only improves the solutions found by MDQL, but also converges faster. MDQL has also been tested on the solutions of dynamic optimization problems. In this paper, it is also shown that the adaptation capabilities observed in MDQL can be improved with IMDQL. IMDQL was tested on the benchmark problems proposed by Jin. Performance evaluation was made using the Collective Mean Fitness metric proposed by Morrison. IMDQL was compared against an standard evolution strategy with the covariance matrix adaptation (CMA-ES) with very promising results.\",\"PeriodicalId\":296192,\"journal\":{\"name\":\"2007 Sixth Mexican International Conference on Artificial Intelligence, Special Session (MICAI)\",\"volume\":\"88 4 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2007-11-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2007 Sixth Mexican International Conference on Artificial Intelligence, Special Session (MICAI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/MICAI.2007.47\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2007 Sixth Mexican International Conference on Artificial Intelligence, Special Session (MICAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MICAI.2007.47","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

摘要

MDQL是一种基于强化学习的算法,用于解决多目标优化问题,已经在几个应用程序中进行了测试,结果很有希望。MDQL将决策变量离散为一组状态,每个状态都与将代理移动到连续状态的操作相关联。一组智能体探索这个状态空间,并能够应用分布式强化学习算法找到帕累托集。帕累托解的精度取决于所选择的状态粒度。更细的状态粒度创建更精确的解决方案,但代价是更大的搜索空间,因此需要更多的计算资源。本文对原来的MDQL算法进行了一个非常重要的改进,以逐步改进Pareto解。这个名为IMDQL的新算法从粗粒度开始寻找初始帕累托集。在此精炼状态空间中建立了每个Pareto解的邻域,并建立了一个新的Pareto集。这个过程一直持续,直到在一个小的阈值内没有更多的改进。结果表明,IMDQL不仅改进了MDQL找到的解,而且收敛速度更快。MDQL还对动态优化问题的解决方案进行了测试。本文还表明,在MDQL中观察到的自适应能力可以通过IMDQL得到改进。IMDQL在Jin提出的基准问题上进行了测试。采用Morrison提出的集体平均适应度(Collective Mean Fitness)指标进行绩效评价。将IMDQL与具有协方差矩阵自适应(CMA-ES)的标准进化策略进行了比较,得到了很好的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Incremental Refinement of Solutions for Dynamic Multi Objective Optimization Problems
MDQL is an algorithm, based on reinforcement learning, for solving multiple objective optimization problems, that has been tested on several applications with promising results. MDQL discretizes the decision variables into a set of states, each associated with actions to move agents to contiguous states. A group of agents explore this state space and are able to find Pareto sets applying a distributed reinforcement learning algorithm. The precision of the Pareto solutions depends on the chosen granularity of the states. A finer granularity on the states creates more precise solutions but at the expense of a larger search space, and consequently the need for more computational resources. In this paper, a very important improvement is introduced into the original MDQL algorithm to incrementally refined the Pareto solutions. The new algorithm, called IMDQL, starts with a coarse granularity to find an initial Pareto set. A vicinity for each of the Pareto solutions in refined and a new Pareto set is founded in this refined state space. This process continues until there is no more improvement within a small threshold value. It is shown that IMDQL not only improves the solutions found by MDQL, but also converges faster. MDQL has also been tested on the solutions of dynamic optimization problems. In this paper, it is also shown that the adaptation capabilities observed in MDQL can be improved with IMDQL. IMDQL was tested on the benchmark problems proposed by Jin. Performance evaluation was made using the Collective Mean Fitness metric proposed by Morrison. IMDQL was compared against an standard evolution strategy with the covariance matrix adaptation (CMA-ES) with very promising results.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Machine Learning Tools to Time Series Forecasting Algorithm for Affective Pattern Recognition by Means of Use of First Initial Momentum Uncertain Reasoning in Multi-agent Ontology Mapping on the Semantic Web Segmentation and Extraction of Morphologic Features from Capillary Images An Intelligent Agent Using a Q-Learning Method to Allocate Replicated Data in a Distributed Database
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1