Caifeng Chen , Qunfeng Liu , Yunpeng Jing , Mingming Zhang , Shi Cheng , Yun Li
{"title":"On the representativeness metric of benchmark problems in numerical optimization","authors":"Caifeng Chen , Qunfeng Liu , Yunpeng Jing , Mingming Zhang , Shi Cheng , Yun Li","doi":"10.1016/j.swevo.2024.101716","DOIUrl":null,"url":null,"abstract":"<div><p>Numerical comparison on benchmark problems is often necessary in evaluating optimization algorithms with or without theoretical analysis. An implicit assumption is that the adopted set of benchmark problems is representative. However, to our knowledge, there are few results about how to evaluate the representativeness of a test suite, partly due to the difficulty of this issue. In this paper, we first define three different levels of representativeness, and open up a window for addressing step by step the issue of representativeness-measuring. Then we turn to address the Type-III representativeness-measuring problem, and provide a metric for this problem. To illustrate how to use the proposed metric, the representativeness-measuring problem of benchmark problems for single-objective unconstrained continuous optimization is examined.</p><p>The analysis covers as many as 1141 single-objective unconstrained continuous benchmark problems, primarily focusing on existing benchmark problems. Based on the defined representativeness metric, some classical features and calculations are used to assess the representativeness of the benchmark problems. Assessment results show that most of the benchmark problems of high representativeness are non-separable problems from the CEC and BBOB test suites. We select the top 5% of most representative problems to build a new test suite, providing a more representative and rigorous reference in verifying the overall performance of the optimization algorithms.</p></div>","PeriodicalId":48682,"journal":{"name":"Swarm and Evolutionary Computation","volume":"91 ","pages":"Article 101716"},"PeriodicalIF":8.2000,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Swarm and Evolutionary Computation","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2210650224002542","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Numerical comparison on benchmark problems is often necessary in evaluating optimization algorithms with or without theoretical analysis. An implicit assumption is that the adopted set of benchmark problems is representative. However, to our knowledge, there are few results about how to evaluate the representativeness of a test suite, partly due to the difficulty of this issue. In this paper, we first define three different levels of representativeness, and open up a window for addressing step by step the issue of representativeness-measuring. Then we turn to address the Type-III representativeness-measuring problem, and provide a metric for this problem. To illustrate how to use the proposed metric, the representativeness-measuring problem of benchmark problems for single-objective unconstrained continuous optimization is examined.
The analysis covers as many as 1141 single-objective unconstrained continuous benchmark problems, primarily focusing on existing benchmark problems. Based on the defined representativeness metric, some classical features and calculations are used to assess the representativeness of the benchmark problems. Assessment results show that most of the benchmark problems of high representativeness are non-separable problems from the CEC and BBOB test suites. We select the top 5% of most representative problems to build a new test suite, providing a more representative and rigorous reference in verifying the overall performance of the optimization algorithms.
期刊介绍:
Swarm and Evolutionary Computation is a pioneering peer-reviewed journal focused on the latest research and advancements in nature-inspired intelligent computation using swarm and evolutionary algorithms. It covers theoretical, experimental, and practical aspects of these paradigms and their hybrids, promoting interdisciplinary research. The journal prioritizes the publication of high-quality, original articles that push the boundaries of evolutionary computation and swarm intelligence. Additionally, it welcomes survey papers on current topics and novel applications. Topics of interest include but are not limited to: Genetic Algorithms, and Genetic Programming, Evolution Strategies, and Evolutionary Programming, Differential Evolution, Artificial Immune Systems, Particle Swarms, Ant Colony, Bacterial Foraging, Artificial Bees, Fireflies Algorithm, Harmony Search, Artificial Life, Digital Organisms, Estimation of Distribution Algorithms, Stochastic Diffusion Search, Quantum Computing, Nano Computing, Membrane Computing, Human-centric Computing, Hybridization of Algorithms, Memetic Computing, Autonomic Computing, Self-organizing systems, Combinatorial, Discrete, Binary, Constrained, Multi-objective, Multi-modal, Dynamic, and Large-scale Optimization.