Contact Us | Language: čeština English
Title: | How does the number of objective function evaluations impact our understanding of metaheuristics behavior? | ||||||||||
Author: | Kazíková, Anežka; Pluháček, Michal; Šenkeřík, Roman | ||||||||||
Document type: | Peer-reviewed article (English) | ||||||||||
Source document: | IEEE Access. 2021, vol. 9, p. 44032-44048 | ||||||||||
ISSN: | 2169-3536 (Sherpa/RoMEO, JCR) | ||||||||||
Journal Impact
This chart shows the development of journal-level impact metrics in time
|
|||||||||||
DOI: | https://doi.org/10.1109/ACCESS.2021.3066135 | ||||||||||
Abstract: | Comparing various metaheuristics based on an equal number of objective function evaluations has become standard practice. Many contemporary publications use a specific number of objective function evaluations by the benchmarking sets definitions. Furthermore, many publications deal with the recurrent theme of late stagnation, which may lead to the impression that continuing the optimization process could be a waste of computational capabilities. But is it? Recently, many challenges, issues, and questions have been raised regarding fair comparisons and recommendations towards good practices for benchmarking metaheuristic algorithms. The aim of this work is not to compare the performance of several well-known algorithms but to investigate the issues that can appear in benchmarking and comparisons of metaheuristics performance (no matter what the problem is). This article studies the impact of a higher evaluation number on a selection of metaheuristic algorithms. We examine the effect of a raised evaluation budget on overall performance, mean convergence, and population diversity of selected swarm algorithms and IEEE CEC competition winners. Even though the final impact varies based on current algorithm selection, it may significantly affect the final verdict of metaheuristics comparison. This work has picked an important benchmarking issue and made extensive analysis, resulting in conclusions and possible recommendations for users working with real engineering optimization problems or researching the metaheuristics algorithms. Especially nowadays, when metaheuristic algorithms are used for increasingly complex optimization problems, and meet machine learning in AutoML frameworks, we conclude that the objective function evaluation budget should be considered another vital optimization input variable. | ||||||||||
Full text: | https://ieeexplore.ieee.org/document/9378523 | ||||||||||
Show full item record |