首页 > 最新文献

Evolutionary Computation最新文献

英文 中文
Hyperparameter Control Using Fuzzy Logic: Evolving Policies for Adaptive Fuzzy Particle Swarm Optimization Algorithm. 使用模糊逻辑的超参数控制:自适应模糊粒子群优化算法的演化策略。
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-06-18 DOI: 10.1162/evco_a_00353
Nicolas Roy, Charlotte Beauthier, Alexandre Mayer

Heuristic optimization methods such as Particle Swarm Optimization depend on their parameters to achieve optimal performance on a given class of problems. Some modifications of heuristic algorithms aim at adapting those parameters during the optimization process. We present a novel approach to design such adaptation strategies using continuous fuzzy feedback control. Fuzzy feedback provides a simple interface where probes are sampled in the optimization process and parameters are fed back to the optimizer. The probes are turned into parameters by a fuzzy process optimized beforehand to maximize performance on a training benchmark. Utilizing this framework, we systematically established 127 different Fuzzy Particle Swarm Optimization algorithms featuring a maximum of 7 parameters under fuzzy control. These newly devised algorithms exhibit superior performance compared to both traditional PSO and some of its best parameter control variants. The performance is reported in the single-objective bound-constrained numerical optimization competition of CEC 2020. Additionally, two specific controls, highlighted for their efficacy and dependability, demonstrated commendable performance in real-world scenarios from CEC 2011.

启发式优化方法(如粒子群优化法)取决于其参数,以便在特定问题类别中实现最佳性能。启发式算法的一些改进旨在优化过程中调整这些参数。我们提出了一种利用连续模糊反馈控制来设计此类适应策略的新方法。模糊反馈提供了一个简单的界面,在优化过程中对探测器进行采样,并将参数反馈给优化器。探测结果通过事先优化的模糊过程转化为参数,从而在训练基准上实现性能最大化。利用这一框架,我们系统地建立了 127 种不同的模糊粒子群优化算法,在模糊控制下最多可有 7 个参数。与传统 PSO 及其一些最佳参数控制变体相比,这些新设计的算法表现出更优越的性能。在 CEC 2020 的单目标约束数值优化竞赛中,报告了这些算法的性能。此外,在 CEC 2011 的实际应用场景中,两种特定的控制方法因其有效性和可靠性而备受瞩目,表现值得称赞。
{"title":"Hyperparameter Control Using Fuzzy Logic: Evolving Policies for Adaptive Fuzzy Particle Swarm Optimization Algorithm.","authors":"Nicolas Roy, Charlotte Beauthier, Alexandre Mayer","doi":"10.1162/evco_a_00353","DOIUrl":"https://doi.org/10.1162/evco_a_00353","url":null,"abstract":"<p><p>Heuristic optimization methods such as Particle Swarm Optimization depend on their parameters to achieve optimal performance on a given class of problems. Some modifications of heuristic algorithms aim at adapting those parameters during the optimization process. We present a novel approach to design such adaptation strategies using continuous fuzzy feedback control. Fuzzy feedback provides a simple interface where probes are sampled in the optimization process and parameters are fed back to the optimizer. The probes are turned into parameters by a fuzzy process optimized beforehand to maximize performance on a training benchmark. Utilizing this framework, we systematically established 127 different Fuzzy Particle Swarm Optimization algorithms featuring a maximum of 7 parameters under fuzzy control. These newly devised algorithms exhibit superior performance compared to both traditional PSO and some of its best parameter control variants. The performance is reported in the single-objective bound-constrained numerical optimization competition of CEC 2020. Additionally, two specific controls, highlighted for their efficacy and dependability, demonstrated commendable performance in real-world scenarios from CEC 2011.</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141421719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Large-Scale Multiobjective Evolutionary Algorithm Guided by Low-Dimensional Surrogates of Scalarization Functions. 以低维标度化函数替代物为指导的大规模多目标进化算法
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-06-18 DOI: 10.1162/evco_a_00354
Haoran Gu, Handing Wang, Cheng He, Bo Yuan, Yaochu Jin

Recently, computationally intensive multiobjective optimization problems have been efficiently solved by surrogate-assisted multiobjective evolutionary algorithms. However, most of those algorithms can only handle no more than 200 decision variables. As the number of decision variables increases further, unreliable surrogate models will result in a dramatic deterioration of their performance, which makes large-scale expensive multiobjective optimization challenging. To address this challenge, we develop a large-scale multiobjective evolutionary algorithm guided by low-dimensional surrogate models of scalarization functions. The proposed algorithm (termed LDS-AF) reduces the dimension of the original decision space based on principal component analysis, and then directly approximates the scalarization functions in a decompositionbased multiobjective evolutionary algorithm. With the help of a two-stage modeling strategy and convergence control strategy, LDS-AF can keep a good balance between convergence and diversity, and achieve a promising performance without being trapped in a local optimum prematurely. The experimental results on a set of test instances have demonstrated its superiority over eight state-of-the-art algorithms on multiobjective optimization problems with up to 1000 decision variables using only 500 real function evaluations.

最近,计算密集型多目标优化问题已通过代理辅助多目标进化算法得到有效解决。然而,这些算法大多只能处理不超过 200 个决策变量。随着决策变量数量的进一步增加,不可靠的代用模型将导致其性能急剧下降,从而使大规模昂贵的多目标优化面临挑战。为了应对这一挑战,我们开发了一种以标量化函数的低维代理模型为指导的大规模多目标进化算法。所提出的算法(称为 LDS-AF)基于主成分分析降低了原始决策空间的维度,然后在基于分解的多目标进化算法中直接逼近标量化函数。借助两阶段建模策略和收敛控制策略,LDS-AF 可以在收敛性和多样性之间保持良好的平衡,并在不过早陷入局部最优的情况下取得可喜的性能。在一组测试实例上的实验结果表明,在多达 1000 个决策变量的多目标优化问题上,LDS-AF 只用了 500 次实际函数评估,就优于八种最先进的算法。
{"title":"Large-Scale Multiobjective Evolutionary Algorithm Guided by Low-Dimensional Surrogates of Scalarization Functions.","authors":"Haoran Gu, Handing Wang, Cheng He, Bo Yuan, Yaochu Jin","doi":"10.1162/evco_a_00354","DOIUrl":"https://doi.org/10.1162/evco_a_00354","url":null,"abstract":"<p><p>Recently, computationally intensive multiobjective optimization problems have been efficiently solved by surrogate-assisted multiobjective evolutionary algorithms. However, most of those algorithms can only handle no more than 200 decision variables. As the number of decision variables increases further, unreliable surrogate models will result in a dramatic deterioration of their performance, which makes large-scale expensive multiobjective optimization challenging. To address this challenge, we develop a large-scale multiobjective evolutionary algorithm guided by low-dimensional surrogate models of scalarization functions. The proposed algorithm (termed LDS-AF) reduces the dimension of the original decision space based on principal component analysis, and then directly approximates the scalarization functions in a decompositionbased multiobjective evolutionary algorithm. With the help of a two-stage modeling strategy and convergence control strategy, LDS-AF can keep a good balance between convergence and diversity, and achieve a promising performance without being trapped in a local optimum prematurely. The experimental results on a set of test instances have demonstrated its superiority over eight state-of-the-art algorithms on multiobjective optimization problems with up to 1000 decision variables using only 500 real function evaluations.</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141421720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural Architecture Search Using Covariance Matrix Adaptation Evolution Strategy. 使用协方差矩阵适应进化策略的神经架构搜索
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-06-03 DOI: 10.1162/evco_a_00331
Nilotpal Sinha, Kuan-Wen Chen

Evolution-based neural architecture search methods have shown promising results, but they require high computational resources because these methods involve training each candidate architecture from scratch and then evaluating its fitness, which results in long search time. Covariance Matrix Adaptation Evolution Strategy (CMA-ES) has shown promising results in tuning hyperparameters of neural networks but has not been used for neural architecture search. In this work, we propose a framework called CMANAS which applies the faster convergence property of CMA-ES to the deep neural architecture search problem. Instead of training each individual architecture seperately, we used the accuracy of a trained one shot model (OSM) on the validation data as a prediction of the fitness of the architecture, resulting in reduced search time. We also used an architecture-fitness table (AF table) for keeping a record of the already evaluated architecture, thus further reducing the search time. The architectures are modeled using a normal distribution, which is updated using CMA-ES based on the fitness of the sampled population. Experimentally, CMANAS achieves better results than previous evolution-based methods while reducing the search time significantly. The effectiveness of CMANAS is shown on two different search spaces using four datasets: CIFAR-10, CIFAR-100, ImageNet, and ImageNet16-120. All the results show that CMANAS is a viable alternative to previous evolution-based methods and extends the application of CMA-ES to the deep neural architecture search field.

基于进化的神经架构搜索方法已显示出良好的效果,但这些方法需要大量的计算资源,因为这些方法涉及从头开始训练每个候选架构,然后评估其适合度,从而导致搜索时间过长。协方差矩阵自适应进化策略(CMA-ES)在调整神经网络超参数方面取得了可喜的成果,但尚未用于神经架构搜索。在这项工作中,我们提出了一个名为 CMANAS 的框架,它将 CMA-ES 的快速收敛特性应用于深度神经架构搜索问题。我们没有单独训练每个架构,而是使用在验证数据上训练的单次模型(OSM)的准确性来预测架构的适配性,从而缩短了搜索时间。我们还使用了架构适配性表(AF 表)来保存已评估架构的记录,从而进一步减少了搜索时间。架构采用正态分布建模,并根据采样群体的适配性使用 CMA-ES 对其进行更新。通过实验,CMANAS 比以前基于进化的方法取得了更好的结果,同时大大缩短了搜索时间。CMANAS 在两个不同的搜索空间中使用四个数据集显示了其有效性:CIFAR-10、CIFAR-100、ImageNet 和 ImageNet16-120。所有结果都表明,CMANAS 是以前基于进化的方法的可行替代方案,并将 CMA-ES 的应用扩展到了深度神经架构搜索领域。
{"title":"Neural Architecture Search Using Covariance Matrix Adaptation Evolution Strategy.","authors":"Nilotpal Sinha, Kuan-Wen Chen","doi":"10.1162/evco_a_00331","DOIUrl":"10.1162/evco_a_00331","url":null,"abstract":"<p><p>Evolution-based neural architecture search methods have shown promising results, but they require high computational resources because these methods involve training each candidate architecture from scratch and then evaluating its fitness, which results in long search time. Covariance Matrix Adaptation Evolution Strategy (CMA-ES) has shown promising results in tuning hyperparameters of neural networks but has not been used for neural architecture search. In this work, we propose a framework called CMANAS which applies the faster convergence property of CMA-ES to the deep neural architecture search problem. Instead of training each individual architecture seperately, we used the accuracy of a trained one shot model (OSM) on the validation data as a prediction of the fitness of the architecture, resulting in reduced search time. We also used an architecture-fitness table (AF table) for keeping a record of the already evaluated architecture, thus further reducing the search time. The architectures are modeled using a normal distribution, which is updated using CMA-ES based on the fitness of the sampled population. Experimentally, CMANAS achieves better results than previous evolution-based methods while reducing the search time significantly. The effectiveness of CMANAS is shown on two different search spaces using four datasets: CIFAR-10, CIFAR-100, ImageNet, and ImageNet16-120. All the results show that CMANAS is a viable alternative to previous evolution-based methods and extends the application of CMA-ES to the deep neural architecture search field.</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9424655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On Single-Objective Sub-Graph-Based Mutation for Solving the Bi-Objective Minimum Spanning Tree Problem. 基于子图的单目标突变用于解决双目标最小生成树问题
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-06-03 DOI: 10.1162/evco_a_00335
Jakob Bossek, Christian Grimme

We contribute to the efficient approximation of the Pareto-set for the classical NP-hard multiobjective minimum spanning tree problem (moMST) adopting evolutionary computation. More precisely, by building upon preliminary work, we analyze the neighborhood structure of Pareto-optimal spanning trees and design several highly biased sub-graph-based mutation operators founded on the gained insights. In a nutshell, these operators replace (un)connected sub-trees of candidate solutions with locally optimal sub-trees. The latter (biased) step is realized by applying Kruskal's single-objective MST algorithm to a weighted sum scalarization of a sub-graph. We prove runtime complexity results for the introduced operators and investigate the desirable Pareto-beneficial property. This property states that mutants cannot be dominated by their parent. Moreover, we perform an extensive experimental benchmark study to showcase the operator's practical suitability. Our results confirm that the sub-graph-based operators beat baseline algorithms from the literature even with severely restricted computational budget in terms of function evaluations on four different classes of complete graphs with different shapes of the Pareto-front.

我们为采用进化计算高效逼近经典 NP 难多目标最小生成树问题(moMST)的帕累托集做出了贡献。更确切地说,在前期工作的基础上,我们分析了帕累托最优生成树的邻域结构,并根据所获得的洞察力设计了几种基于子图的高偏置突变算子。简而言之,这些算子用局部最优子树替换候选解决方案中(不)相连的子树。后一步(偏置)是通过将 Kruskal 的单目标 MST 算法应用于子图的加权和标量化来实现的。我们证明了所引入算子的运行时间复杂性结果,并研究了理想的帕累托效益特性。这一特性表明,突变体不会被其父图所支配。此外,我们还进行了广泛的实验基准研究,以展示算子的实际适用性。我们的研究结果证实,基于子图的算子在具有不同帕累托前沿形状的四类不同完整图上的函数评估方面,即使计算预算严重受限,也能击败文献中的基准算法。
{"title":"On Single-Objective Sub-Graph-Based Mutation for Solving the Bi-Objective Minimum Spanning Tree Problem.","authors":"Jakob Bossek, Christian Grimme","doi":"10.1162/evco_a_00335","DOIUrl":"10.1162/evco_a_00335","url":null,"abstract":"<p><p>We contribute to the efficient approximation of the Pareto-set for the classical NP-hard multiobjective minimum spanning tree problem (moMST) adopting evolutionary computation. More precisely, by building upon preliminary work, we analyze the neighborhood structure of Pareto-optimal spanning trees and design several highly biased sub-graph-based mutation operators founded on the gained insights. In a nutshell, these operators replace (un)connected sub-trees of candidate solutions with locally optimal sub-trees. The latter (biased) step is realized by applying Kruskal's single-objective MST algorithm to a weighted sum scalarization of a sub-graph. We prove runtime complexity results for the introduced operators and investigate the desirable Pareto-beneficial property. This property states that mutants cannot be dominated by their parent. Moreover, we perform an extensive experimental benchmark study to showcase the operator's practical suitability. Our results confirm that the sub-graph-based operators beat baseline algorithms from the literature even with severely restricted computational budget in terms of function evaluations on four different classes of complete graphs with different shapes of the Pareto-front.</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9967379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Role of Morphological Variation in Evolutionary Robotics: Maximizing Performance and Robustness. 形态变异在进化机器人学中的作用:最大化性能和鲁棒性
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-06-03 DOI: 10.1162/evco_a_00336
Jonata Tyska Carvalho, Stefano Nolfi

Exposing an evolutionary algorithm that is used to evolve robot controllers to variable conditions is necessary to obtain solutions which are robust and can cross the reality gap. However, we do not yet have methods for analyzing and understanding the impact of the varying morphological conditions which impact the evolutionary process, and therefore for choosing suitable variation ranges. By morphological conditions, we refer to the starting state of the robot, and to variations in its sensor readings during operation due to noise. In this paper, we introduce a method that permits us to measure the impact of these morphological variations and we analyze the relation between the amplitude of variations, the modality with which they are introduced, and the performance and robustness of evolving agents. Our results demonstrate that (i) the evolutionary algorithm can tolerate morphological variations which have a very high impact, (ii) variations affecting the actions of the agent are tolerated much better than variations affecting the initial state of the agent or of the environment, and (iii) improving the accuracy of the fitness measure through multiple evaluations is not always useful. Moreover, our results show that morphological variations permit generating solutions which perform better both in varying and non-varying conditions.

要想获得稳健并能跨越现实鸿沟的解决方案,就必须让用于进化机器人控制器的进化算法面临各种变化条件。然而,我们还没有方法来分析和理解影响进化过程的不同形态条件的影响,从而选择合适的变化范围。所谓形态条件,指的是机器人的起始状态,以及运行过程中由于噪音导致的传感器读数变化。在本文中,我们介绍了一种可以测量这些形态变化影响的方法,并分析了变化幅度、引入变化的方式以及进化代理的性能和鲁棒性之间的关系。我们的结果表明:(i) 进化算法可以容忍影响非常大的形态变化;(ii) 与影响代理初始状态或环境的变化相比,影响代理行动的变化的容忍度要高得多;(iii) 通过多次评估提高适应度测量的准确性并不总是有用的。此外,我们的结果表明,形态变化允许生成在变化和非变化条件下都表现更好的解决方案。
{"title":"The Role of Morphological Variation in Evolutionary Robotics: Maximizing Performance and Robustness.","authors":"Jonata Tyska Carvalho, Stefano Nolfi","doi":"10.1162/evco_a_00336","DOIUrl":"10.1162/evco_a_00336","url":null,"abstract":"<p><p>Exposing an evolutionary algorithm that is used to evolve robot controllers to variable conditions is necessary to obtain solutions which are robust and can cross the reality gap. However, we do not yet have methods for analyzing and understanding the impact of the varying morphological conditions which impact the evolutionary process, and therefore for choosing suitable variation ranges. By morphological conditions, we refer to the starting state of the robot, and to variations in its sensor readings during operation due to noise. In this paper, we introduce a method that permits us to measure the impact of these morphological variations and we analyze the relation between the amplitude of variations, the modality with which they are introduced, and the performance and robustness of evolving agents. Our results demonstrate that (i) the evolutionary algorithm can tolerate morphological variations which have a very high impact, (ii) variations affecting the actions of the agent are tolerated much better than variations affecting the initial state of the agent or of the environment, and (iii) improving the accuracy of the fitness measure through multiple evaluations is not always useful. Moreover, our results show that morphological variations permit generating solutions which perform better both in varying and non-varying conditions.</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9726876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparing Robot Controller Optimization Methods on Evolvable Morphologies. 比较可进化形态上的机器人控制器优化方法
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-06-03 DOI: 10.1162/evco_a_00334
Fuda van Diggelen, Eliseo Ferrante, A E Eiben

In this paper, we compare Bayesian Optimization, Differential Evolution, and an Evolution Strategy employed as a gait-learning algorithm in modular robots. The motivational scenario is the joint evolution of morphologies and controllers, where "newborn" robots also undergo a learning process to optimize their inherited controllers (without changing their bodies). This context raises the question: How do gait-learning algorithms compare when applied to various morphologies that are not known in advance (and thus need to be treated as without priors)? To answer this question, we use a test suite of twenty different robot morphologies to evaluate our gait-learners and compare their efficiency, efficacy, and sensitivity to morphological differences. The results indicate that Bayesian Optimization and Differential Evolution deliver the same solution quality (walking speed for the robot) with fewer evaluations than the Evolution Strategy. Furthermore, the Evolution Strategy is more sensitive for morphological differences (its efficacy varies more between different morphologies) and is more subject to luck (repeated runs on the same morphology show greater variance in the outcomes).

在本文中,我们比较了贝叶斯优化法、差分进化法和一种在模块化机器人中用作步态学习算法的进化策略。本文的动机是形态和控制器的联合进化,其中 "新生 "机器人也经历了一个学习过程,以优化其继承的控制器(而不改变其身体)。在这种情况下,就产生了一个问题:当步态学习算法应用于事先未知的各种形态(因此需要被视为无先验)时,它们之间的比较如何?为了回答这个问题,我们使用了一个包含二十种不同机器人形态的测试套件来评估我们的步态学习算法,并比较它们的效率、功效以及对形态差异的敏感性。结果表明,与进化策略相比,贝叶斯优化和差分进化能以更少的评估次数提供相同质量的解决方案(机器人的行走速度)。此外,进化策略对形态差异更为敏感(其功效在不同形态之间的差异更大),而且更容易受到运气的影响(在同一形态上重复运行的结果差异更大)。
{"title":"Comparing Robot Controller Optimization Methods on Evolvable Morphologies.","authors":"Fuda van Diggelen, Eliseo Ferrante, A E Eiben","doi":"10.1162/evco_a_00334","DOIUrl":"10.1162/evco_a_00334","url":null,"abstract":"<p><p>In this paper, we compare Bayesian Optimization, Differential Evolution, and an Evolution Strategy employed as a gait-learning algorithm in modular robots. The motivational scenario is the joint evolution of morphologies and controllers, where \"newborn\" robots also undergo a learning process to optimize their inherited controllers (without changing their bodies). This context raises the question: How do gait-learning algorithms compare when applied to various morphologies that are not known in advance (and thus need to be treated as without priors)? To answer this question, we use a test suite of twenty different robot morphologies to evaluate our gait-learners and compare their efficiency, efficacy, and sensitivity to morphological differences. The results indicate that Bayesian Optimization and Differential Evolution deliver the same solution quality (walking speed for the robot) with fewer evaluations than the Evolution Strategy. Furthermore, the Evolution Strategy is more sensitive for morphological differences (its efficacy varies more between different morphologies) and is more subject to luck (repeated runs on the same morphology show greater variance in the outcomes).</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9541798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Virtual Position Guided Strategy for Particle Swarm Optimization Algorithms on Multimodal Problems. 多模态问题上粒子群优化算法的虚拟位置引导策略
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-05-21 DOI: 10.1162/evco_a_00352
Chao Li, Jun Sun, Li-Wei Li, Min Shan, Vasile Palade, Xiaojun Wu

Premature convergence is a thorny problem for particle swarm optimization (PSO) algorithms, especially on multimodal problems, where maintaining swarm diversity is crucial. However, most enhancement strategies for PSO, including the existing diversity-guided strategies, have not fully addressed this issue. This paper proposes the virtual position guided (VPG) strategy for PSO algorithms. The VPG strategy calculates diversity values for two different populations and establishes a diversity baseline. It then dynamically guides the algorithm to conduct different search behaviors, through three phases - divergence, normal, and acceleration - in each iteration, based on the relationships among these diversity values and the baseline. Collectively, these phases orchestrate different schemes to balance exploration and exploitation, collaboratively steering the algorithm away from local optima and towards enhanced solution quality. The introduction of 'virtual position' caters to the strategy's adaptability across various PSO algorithms, ensuring the generality and effectiveness of the proposed VPG strategy. With a single hyperparameter and a recommended usual setup, VPG is easy to implement. The experimental results demonstrate that the VPG strategy is superior to several canonical and the state-of-the-art strategies for diversity guidance, and is effective in improving the search performance of most PSO algorithms on multimodal problems of various dimensionalities.

对于粒子群优化(PSO)算法来说,过早收敛是一个棘手的问题,尤其是在多模式问题上,保持粒子群的多样性至关重要。然而,大多数 PSO 增强策略,包括现有的多样性引导策略,都没有完全解决这个问题。本文提出了 PSO 算法的虚拟位置引导(VPG)策略。VPG 策略计算两个不同种群的多样性值,并建立多样性基线。然后,它根据这些多样性值和基线之间的关系,通过发散、正常和加速三个阶段,在每次迭代中动态指导算法进行不同的搜索行为。这些阶段共同协调不同的方案,以平衡探索和利用,共同引导算法远离局部最优,提高解决方案的质量。虚拟位置 "的引入满足了该策略对各种 PSO 算法的适应性,确保了所提出的 VPG 策略的通用性和有效性。只需一个超参数和推荐的常规设置,VPG 即可轻松实现。实验结果表明,VPG 策略优于几种典型策略和最先进的多样性引导策略,并能有效提高大多数 PSO 算法在不同维度的多模态问题上的搜索性能。
{"title":"Virtual Position Guided Strategy for Particle Swarm Optimization Algorithms on Multimodal Problems.","authors":"Chao Li, Jun Sun, Li-Wei Li, Min Shan, Vasile Palade, Xiaojun Wu","doi":"10.1162/evco_a_00352","DOIUrl":"https://doi.org/10.1162/evco_a_00352","url":null,"abstract":"<p><p>Premature convergence is a thorny problem for particle swarm optimization (PSO) algorithms, especially on multimodal problems, where maintaining swarm diversity is crucial. However, most enhancement strategies for PSO, including the existing diversity-guided strategies, have not fully addressed this issue. This paper proposes the virtual position guided (VPG) strategy for PSO algorithms. The VPG strategy calculates diversity values for two different populations and establishes a diversity baseline. It then dynamically guides the algorithm to conduct different search behaviors, through three phases - divergence, normal, and acceleration - in each iteration, based on the relationships among these diversity values and the baseline. Collectively, these phases orchestrate different schemes to balance exploration and exploitation, collaboratively steering the algorithm away from local optima and towards enhanced solution quality. The introduction of 'virtual position' caters to the strategy's adaptability across various PSO algorithms, ensuring the generality and effectiveness of the proposed VPG strategy. With a single hyperparameter and a recommended usual setup, VPG is easy to implement. The experimental results demonstrate that the VPG strategy is superior to several canonical and the state-of-the-art strategies for diversity guidance, and is effective in improving the search performance of most PSO algorithms on multimodal problems of various dimensionalities.</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141082836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthesising Diverse and Discriminatory Sets of Instances using Novelty Search in Combinatorial Domains. 在组合领域利用新颖性搜索合成多样化和辨别性实例集。
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-05-06 DOI: 10.1162/evco_a_00350
Alejandro Marrero, Eduardo Segredo, Coromoto León, Emma Hart

Gathering sufficient instance data to either train algorithm-selection models or understand algorithm footprints within an instance space can be challenging. We propose an approach to generating synthetic instances that are tailored to perform well with respect to a target algorithm belonging to a predefined portfolio but are also diverse with respect to their features. Our approach uses a novelty search algorithm with a linearly weighted fitness function that balances novelty and performance to generate a large set of diverse and discriminatory instances in a single run of the algorithm. We consider two definitions of novelty: (1) with respect to discriminatory performance within a portfolio of solvers; (2) with respect to the features of the evolved instances. We evaluate the proposed method with respect to its ability to generate diverse and discriminatory instances in two domains (knapsack and bin-packing), comparing to another well-known quality diversity method, Multi-dimensional Archive of Phenotypic Elites (MAP-Elites) and an evolutionary algorithm that only evolves for discriminatory behaviour. The results demonstrate that the novelty search method outperforms its competitors in terms of coverage of the space and its ability to generate instances that are diverse regarding the relative size of the "performance gap" between the target solver and the remaining solvers in the portfolio. Moreover, for the Knapsack domain, we also show that we are able to generate novel instances in regions of an instance space not covered by existing benchmarks using a portfolio of state-of-the-art solvers. Finally, we demonstrate that the method is robust to different portfolios of solvers (stochastic approaches, deterministic heuristics and state-of-the-art methods), thereby providing further evidence of its generality.

为训练算法选择模型或了解算法在实例空间中的足迹而收集足够的实例数据是一项挑战。我们提出了一种生成合成实例的方法,这些实例经过定制,在属于预定义组合的目标算法方面表现良好,但在特征方面也具有多样性。我们的方法使用一种新颖性搜索算法,其线性加权适配函数可在新颖性和性能之间取得平衡,从而在算法的单次运行中生成大量具有多样性和鉴别性的实例。我们考虑了新颖性的两种定义:(1) 与求解器组合中的判别性能有关;(2) 与演化实例的特征有关。我们评估了所提出的方法在两个领域(knapsack 和 bin-packing)中生成多样化和辨别性实例的能力,并将其与另一种著名的质量多样化方法--表型精英多维档案(MAP-Elites)和一种只为辨别行为而进化的进化算法进行了比较。结果表明,新颖性搜索方法在空间覆盖率和生成实例的能力方面优于其竞争对手,而在目标求解器与组合中其余求解器之间 "性能差距 "的相对大小方面,新颖性搜索方法也具有多样性。此外,对于 Knapsack 领域,我们还证明了我们能够使用最先进的求解器组合,在现有基准未覆盖的实例空间区域生成新实例。最后,我们证明了该方法对不同求解器组合(随机方法、确定性启发式方法和最先进方法)的鲁棒性,从而进一步证明了该方法的通用性。
{"title":"Synthesising Diverse and Discriminatory Sets of Instances using Novelty Search in Combinatorial Domains.","authors":"Alejandro Marrero, Eduardo Segredo, Coromoto León, Emma Hart","doi":"10.1162/evco_a_00350","DOIUrl":"https://doi.org/10.1162/evco_a_00350","url":null,"abstract":"<p><p>Gathering sufficient instance data to either train algorithm-selection models or understand algorithm footprints within an instance space can be challenging. We propose an approach to generating synthetic instances that are tailored to perform well with respect to a target algorithm belonging to a predefined portfolio but are also diverse with respect to their features. Our approach uses a novelty search algorithm with a linearly weighted fitness function that balances novelty and performance to generate a large set of diverse and discriminatory instances in a single run of the algorithm. We consider two definitions of novelty: (1) with respect to discriminatory performance within a portfolio of solvers; (2) with respect to the features of the evolved instances. We evaluate the proposed method with respect to its ability to generate diverse and discriminatory instances in two domains (knapsack and bin-packing), comparing to another well-known quality diversity method, Multi-dimensional Archive of Phenotypic Elites (MAP-Elites) and an evolutionary algorithm that only evolves for discriminatory behaviour. The results demonstrate that the novelty search method outperforms its competitors in terms of coverage of the space and its ability to generate instances that are diverse regarding the relative size of the \"performance gap\" between the target solver and the remaining solvers in the portfolio. Moreover, for the Knapsack domain, we also show that we are able to generate novel instances in regions of an instance space not covered by existing benchmarks using a portfolio of state-of-the-art solvers. Finally, we demonstrate that the method is robust to different portfolios of solvers (stochastic approaches, deterministic heuristics and state-of-the-art methods), thereby providing further evidence of its generality.</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140877841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Layered Learning Approach to Scaling in Learning Classifier Systems for Boolean Problems. 布尔问题分类器学习系统中的分层学习扩展方法。
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-05-06 DOI: 10.1162/evco_a_00351
Isidro M Alvarez, Trung B Nguyen, Will N Browne, Mengjie Zhang

Evolutionary Computation (EC) often throws away learned knowledge as it is reset for each new problem addressed. Conversely, humans can learn from small-scale problems, retain this knowledge (plus functionality) and then successfully reuse them in larger-scale and/or related problems. Linking solutions to problems together has been achieved through layered learning, where an experimenter sets a series of simpler related problems to solve a more complex task. Recent works on Learning Classifier Systems (LCSs) has shown that knowledge reuse through the adoption of Code Fragments, GP-like tree-based programs, is plausible. However, random reuse is inefficient. Thus, the research question is how LCS can adopt a layered-learning framework, such that increasingly complex problems can be solved efficiently? An LCS (named XCSCF*) has been developed to include the required base axioms necessary for learning, refined methods for transfer learning and learning recast as a decomposition into a series of subordinate problems. These subordinate problems can be set as a curriculum by a teacher, but this does not mean that an agent can learn from it. Especially if it only extracts over-fitted knowledge of each problem rather than the underlying scalable patterns and functions. Results show that from a conventional tabula rasa, with only a vague notion of what subordinate problems might be relevant, XCSCF* captures the general logic behind the tested domains and therefore can solve any n-bit Multiplexer, n-bit Carry-one, n-bit Majority-on, and n-bit Even-parity problems. This work demonstrates a step towards continual learning as learned knowledge is effectively reused in subsequent problems.

进化计算(EC)通常会丢弃已学知识,因为每解决一个新问题,都要重新设置这些知识。相反,人类可以从小规模的问题中学习,保留这些知识(以及功能),然后成功地在更大规模和/或相关的问题中重复使用。通过分层学习,实验者可以设置一系列较简单的相关问题来解决较复杂的任务,从而将问题的解决方案联系在一起。最近关于学习分类器系统(LCS)的研究表明,通过采用代码片段(类似于 GP 的树状程序)进行知识重用是可行的。然而,随机重用的效率很低。因此,研究的问题是学习分类系统如何采用分层学习框架,从而高效地解决日益复杂的问题?我们开发了一种 LCS(名为 XCSCF*),其中包括学习所需的基本公理、迁移学习的精炼方法以及分解为一系列下级问题的学习重构。这些下属问题可以由教师设置为课程,但这并不意味着代理可以从中学习。特别是如果它只是提取每个问题的过度拟合知识,而不是潜在的可扩展模式和函数。结果表明,XCSCF*能从传统的表格中捕捉到测试领域背后的一般逻辑,因此能解决任何n位多路复用器、n位携带一、n位多数开和n位偶奇偶问题。这项工作展示了向持续学习迈出的一步,因为学到的知识可以在后续问题中有效地重复使用。
{"title":"A Layered Learning Approach to Scaling in Learning Classifier Systems for Boolean Problems.","authors":"Isidro M Alvarez, Trung B Nguyen, Will N Browne, Mengjie Zhang","doi":"10.1162/evco_a_00351","DOIUrl":"https://doi.org/10.1162/evco_a_00351","url":null,"abstract":"<p><p>Evolutionary Computation (EC) often throws away learned knowledge as it is reset for each new problem addressed. Conversely, humans can learn from small-scale problems, retain this knowledge (plus functionality) and then successfully reuse them in larger-scale and/or related problems. Linking solutions to problems together has been achieved through layered learning, where an experimenter sets a series of simpler related problems to solve a more complex task. Recent works on Learning Classifier Systems (LCSs) has shown that knowledge reuse through the adoption of Code Fragments, GP-like tree-based programs, is plausible. However, random reuse is inefficient. Thus, the research question is how LCS can adopt a layered-learning framework, such that increasingly complex problems can be solved efficiently? An LCS (named XCSCF*) has been developed to include the required base axioms necessary for learning, refined methods for transfer learning and learning recast as a decomposition into a series of subordinate problems. These subordinate problems can be set as a curriculum by a teacher, but this does not mean that an agent can learn from it. Especially if it only extracts over-fitted knowledge of each problem rather than the underlying scalable patterns and functions. Results show that from a conventional tabula rasa, with only a vague notion of what subordinate problems might be relevant, XCSCF* captures the general logic behind the tested domains and therefore can solve any n-bit Multiplexer, n-bit Carry-one, n-bit Majority-on, and n-bit Even-parity problems. This work demonstrates a step towards continual learning as learned knowledge is effectively reused in subsequent problems.</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140877840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
OneMax is not the Easiest Function for Fitness Improvements. OneMax 并非改善体能的最简单功能。
IF 6.8 2区 计算机科学 Q1 Mathematics Pub Date : 2024-03-22 DOI: 10.1162/evco_a_00348
Marc Kaufmann, Maxime Larcher, Johannes Lengler, Xun Zou

We study the (1:s+1) success rule for controlling the population size of the (1,λ)- EA. It was shown by Hevia Fajardo and Sudholt that this parameter control mechanism can run into problems for large s if the fitness landscape is too easy. They conjectured that this problem is worst for the ONEMAX benchmark, since in some well-established sense ONEMAX is known to be the easiest fitness landscape. In this paper we disprove this conjecture. We show that there exist s and ɛ such that the self-adjusting (1,λ)-EA with the (1:s+1)-rule optimizes ONEMAX efficiently when started with ɛn zero-bits, but does not find the optimum in polynomial time on DYNAMIC BINVAL. Hence, we show that there are landscapes where the problem of the (1:s+1)-rule for controlling the population size of the (1,λ)-EA is more severe than for ONEMAX. The key insight is that, while ONEMAX is the easiest function for decreasing the distance to the optimum, it is not the easiest fitness landscape with respect to finding fitness-improving steps.

我们研究了控制 (1,λ)- EA 种群规模的 (1:s+1) 成功规则。Hevia Fajardo 和 Sudholt 的研究表明,如果适配景观过于简单,这种参数控制机制在 s 较大时可能会出现问题。他们推测,这个问题在 ONEMAX 基准中最为严重,因为从某种既定的意义上讲,ONEMAX 是已知的最简单的适配景观。在本文中,我们推翻了这一猜想。我们证明,存在 s 和 ɛ 这样的情况:采用 (1:s+1) 规则的自调整 (1,λ)-EA 在从ɛn 个零位开始时能高效优化 ONEMAX,但在动态 BINVAL 上却不能在多项式时间内找到最优。因此,我们证明,在有些地形中,控制 (1,λ)-EA 种群规模的 (1:s+1)- 规则的问题比 ONEMAX 更严重。关键之处在于,虽然ONEMAX 是最容易减小与最优值距离的函数,但它并不是最容易找到改善适应性步骤的适应性景观。
{"title":"OneMax is not the Easiest Function for Fitness Improvements.","authors":"Marc Kaufmann, Maxime Larcher, Johannes Lengler, Xun Zou","doi":"10.1162/evco_a_00348","DOIUrl":"https://doi.org/10.1162/evco_a_00348","url":null,"abstract":"<p><p>We study the (1:s+1) success rule for controlling the population size of the (1,λ)- EA. It was shown by Hevia Fajardo and Sudholt that this parameter control mechanism can run into problems for large s if the fitness landscape is too easy. They conjectured that this problem is worst for the ONEMAX benchmark, since in some well-established sense ONEMAX is known to be the easiest fitness landscape. In this paper we disprove this conjecture. We show that there exist s and ɛ such that the self-adjusting (1,λ)-EA with the (1:s+1)-rule optimizes ONEMAX efficiently when started with ɛn zero-bits, but does not find the optimum in polynomial time on DYNAMIC BINVAL. Hence, we show that there are landscapes where the problem of the (1:s+1)-rule for controlling the population size of the (1,λ)-EA is more severe than for ONEMAX. The key insight is that, while ONEMAX is the easiest function for decreasing the distance to the optimum, it is not the easiest fitness landscape with respect to finding fitness-improving steps.</p>","PeriodicalId":50470,"journal":{"name":"Evolutionary Computation","volume":null,"pages":null},"PeriodicalIF":6.8,"publicationDate":"2024-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140295208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Evolutionary Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1