Dr Lee Christie l.a.christie@rgu.ac.uk
Research Fellow
Investigating benchmark correlations when comparing algorithms with parameter tuning: detailed experiments and results.
Christie, Lee A.; Brownlee, Alexander E.I.; Woodward, John R.
Authors
Alexander E.I. Brownlee
John R. Woodward
Abstract
Benchmarks are important to demonstrate the utility of optimisation algorithms, but there is controversy about the practice of benchmarking; we could select instances that present our algorithm favourably, and dismiss those on which our algorithm underperforms. Several papers highlight the pitfalls concerned with benchmarking, some of which concern the context of the automated design of algorithms, where we use a set of problem instances (benchmarks) to train our algorithm. As with machine learning, if the training set does not reflect the test set, the algorithm will not generalize. This raises some open questions concerning the use of test instances to automatically design algorithms. We use differential evolution and sweep the parameter settings to investigate the practice of benchmarking using the BBOB benchmarks. We make three key findings. Firstly, several benchmark functions are highly correlated. This may lead to the false conclusion that an algorithm performs well in general, when it performs poorly on a few key instances, possibly introducing unwanted bias to a resulting automatically designed algorithm. Secondly, the number of evaluations can have a large effect on the conclusion. Finally, a systematic sweep of the parameters shows how performance varies with time across the space of algorithm configurations. The datasets, including all computed features, the evolved policies and their performances, and the visualisations for all feature sets are available from the University of Stirling Data Repository (http://hdl.handle.net/11667/109).
Citation
CHRISTIE, L.A., BROWNLEE, A.E.I. and WOODWARD, J.R. 2018. Investigating benchmark correlations when comparing algorithms with parameter tuning: detailed experiments and results. Stirling: University of Stirling [online]. Available from: http://hdl.handle.net/1893/26956
Report Type | Technical Report |
---|---|
Online Publication Date | Apr 30, 2018 |
Publication Date | Apr 30, 2018 |
Deposit Date | Mar 9, 2020 |
Publicly Available Date | Mar 25, 2020 |
Publisher | University of Stirling |
Keywords | Benchmarking; Benchmarks; Black-Box Optimization Benchmarking (BBOB); Differential evolution; Continuous optimisation; Algorithm automation |
Public URL | https://rgu-repository.worktribe.com/output/876294 |
Publisher URL | http://hdl.handle.net/1893/26956 |
Related Public URLs | https://rgu-repository.worktribe.com/output/876279 |
Files
CHRISTIE 2018 Investigating benchmark correlations (REPORT)
(1.1 Mb)
PDF
Publisher Licence URL
https://creativecommons.org/licenses/by-nc/4.0/
You might also like
Multi-objective evolutionary design of antibiotic treatments.
(2019)
Journal Article
Investigating benchmark correlations when comparing algorithms with parameter tuning.
(2018)
Presentation / Conference Contribution
Downloadable Citations
About OpenAIR@RGU
Administrator e-mail: publications@rgu.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search