Vinícius Veloso de Melo, Alexandre Moreira Nascimento, Giovanni Iacca
{"title":"用于约束优化的具有自适应惩罚函数的协同进化算法","authors":"Vinícius Veloso de Melo, Alexandre Moreira Nascimento, Giovanni Iacca","doi":"10.1007/s00500-024-09896-5","DOIUrl":null,"url":null,"abstract":"<p>Several constrained optimization problems have been adequately solved over the years thanks to the advances in the area of metaheuristics. Nevertheless, the question as to which search logic performs better on constrained optimization often arises. In this paper, we present Dual Search Optimization (DSO), a co-evolutionary algorithm that includes an adaptive penalty function to handle constrained problems. Compared to other self-adaptive metaheuristics, one of the main advantages of DSO is that it is able auto-construct its own perturbation logics, i.e., the ways solutions are modified to create new ones during the optimization process. This is accomplished by co-evolving the solutions (encoded as vectors of integer/real values) and perturbation strategies (encoded as Genetic Programming trees), in order to adapt the search to the problem. In addition to that, the adaptive penalty function allows the algorithm to handle constraints very effectively, yet with a minor additional algorithmic overhead. We compare DSO with several algorithms from the state-of-the-art on two sets of problems, namely: (1) seven well-known constrained engineering design problems and (2) the CEC 2017 benchmark for constrained optimization. Our results show that DSO can achieve state-of-the-art performances, being capable to automatically adjust its behavior to the problem at hand.</p>","PeriodicalId":22039,"journal":{"name":"Soft Computing","volume":"86 1","pages":""},"PeriodicalIF":3.1000,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A co-evolutionary algorithm with adaptive penalty function for constrained optimization\",\"authors\":\"Vinícius Veloso de Melo, Alexandre Moreira Nascimento, Giovanni Iacca\",\"doi\":\"10.1007/s00500-024-09896-5\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Several constrained optimization problems have been adequately solved over the years thanks to the advances in the area of metaheuristics. Nevertheless, the question as to which search logic performs better on constrained optimization often arises. In this paper, we present Dual Search Optimization (DSO), a co-evolutionary algorithm that includes an adaptive penalty function to handle constrained problems. Compared to other self-adaptive metaheuristics, one of the main advantages of DSO is that it is able auto-construct its own perturbation logics, i.e., the ways solutions are modified to create new ones during the optimization process. This is accomplished by co-evolving the solutions (encoded as vectors of integer/real values) and perturbation strategies (encoded as Genetic Programming trees), in order to adapt the search to the problem. In addition to that, the adaptive penalty function allows the algorithm to handle constraints very effectively, yet with a minor additional algorithmic overhead. We compare DSO with several algorithms from the state-of-the-art on two sets of problems, namely: (1) seven well-known constrained engineering design problems and (2) the CEC 2017 benchmark for constrained optimization. Our results show that DSO can achieve state-of-the-art performances, being capable to automatically adjust its behavior to the problem at hand.</p>\",\"PeriodicalId\":22039,\"journal\":{\"name\":\"Soft Computing\",\"volume\":\"86 1\",\"pages\":\"\"},\"PeriodicalIF\":3.1000,\"publicationDate\":\"2024-07-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Soft Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.1007/s00500-024-09896-5\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s00500-024-09896-5","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
A co-evolutionary algorithm with adaptive penalty function for constrained optimization
Several constrained optimization problems have been adequately solved over the years thanks to the advances in the area of metaheuristics. Nevertheless, the question as to which search logic performs better on constrained optimization often arises. In this paper, we present Dual Search Optimization (DSO), a co-evolutionary algorithm that includes an adaptive penalty function to handle constrained problems. Compared to other self-adaptive metaheuristics, one of the main advantages of DSO is that it is able auto-construct its own perturbation logics, i.e., the ways solutions are modified to create new ones during the optimization process. This is accomplished by co-evolving the solutions (encoded as vectors of integer/real values) and perturbation strategies (encoded as Genetic Programming trees), in order to adapt the search to the problem. In addition to that, the adaptive penalty function allows the algorithm to handle constraints very effectively, yet with a minor additional algorithmic overhead. We compare DSO with several algorithms from the state-of-the-art on two sets of problems, namely: (1) seven well-known constrained engineering design problems and (2) the CEC 2017 benchmark for constrained optimization. Our results show that DSO can achieve state-of-the-art performances, being capable to automatically adjust its behavior to the problem at hand.
期刊介绍:
Soft Computing is dedicated to system solutions based on soft computing techniques. It provides rapid dissemination of important results in soft computing technologies, a fusion of research in evolutionary algorithms and genetic programming, neural science and neural net systems, fuzzy set theory and fuzzy systems, and chaos theory and chaotic systems.
Soft Computing encourages the integration of soft computing techniques and tools into both everyday and advanced applications. By linking the ideas and techniques of soft computing with other disciplines, the journal serves as a unifying platform that fosters comparisons, extensions, and new applications. As a result, the journal is an international forum for all scientists and engineers engaged in research and development in this fast growing field.