Hyper-Parameter Tuning for an evolutionary algorithm
Abstract: In this talk, I will present a case study to illustrate how automated algorithm configuration can be used to gain insights into theoretical results on an evolutionary algorithm, namely the (1 (λ,λ)) Genetic Algorithm. This work is a collaboration with Carola Doerr.
The (1 (λ,λ)) Genetic Algorithm is an evolutionary algorithm that has interesting theoretical properties. It is the first algorithm where the benefit of crossover operator is rigorously proved. It is also the first example where self-adjusting parameter choice is proved to outperform any static parameter choice. However, it is not very well understood how the hyper-parameter settings influences the overall performance of the algorithm. Analyzing such multi-dimensional dependencies precisely is at the edge of what running time analysis can offer. In this work, we make a step forward on this question by presenting an in-depth study of the algorithm’s hyper-parameters using techniques in automated algorithm configuration.
Speaker bio: Dr Nguyen Dang is a post-doc in the Constraint Programming group at the University of St Andrews. Her main research focus is on automated algorithm configuration, algorithm selection and their applications in various contexts. These techniques make use of statistical methods and machine learning for fine-tuning of algorithm parameters, assessing parameters’ importance and building algorithm portfolios. Another line of her research is about solving combinatorial optimisation problems using metaheuristic algorithms.