Optimization of black-box functions has been of interest to researchers for many years and has become more relevant since the advent of reinforcement learning problems, which goes along with the development of machine learning. One of the ways used to tackle the problem is the use of evolutionary strategy algorithms. These are able to optimize the given function without the need to compute the gradient of the function itself, which is the main problem while dealing with black-box functions, and they also have theoretical guarantees for their ability to converge to an optimum. After a brief discussion of state-of-the-art algorithms, in this thesis a novel algorithm is presented and compared to them. The algorithm, called ASHGF, implements adaptivity of parameters to escape local global minima and use of historical gradients in order to deal with the exploration-exploitation trade-off. Some theoretical results are given, and it is further numerically validated. All the algorithms are first compared on a standard testbed and then on a set of reinforcement learning problems. It will be shown that the algorithm can outperform all the other state-of-the-art algorithms.

Adaptive evolution strategies for black box optimization

Manente, Alessandro
2021/2022

Abstract

Optimization of black-box functions has been of interest to researchers for many years and has become more relevant since the advent of reinforcement learning problems, which goes along with the development of machine learning. One of the ways used to tackle the problem is the use of evolutionary strategy algorithms. These are able to optimize the given function without the need to compute the gradient of the function itself, which is the main problem while dealing with black-box functions, and they also have theoretical guarantees for their ability to converge to an optimum. After a brief discussion of state-of-the-art algorithms, in this thesis a novel algorithm is presented and compared to them. The algorithm, called ASHGF, implements adaptivity of parameters to escape local global minima and use of historical gradients in order to deal with the exploration-exploitation trade-off. Some theoretical results are given, and it is further numerically validated. All the algorithms are first compared on a standard testbed and then on a set of reinforcement learning problems. It will be shown that the algorithm can outperform all the other state-of-the-art algorithms.
2021-09-22
69
evolution strategies
File in questo prodotto:
File Dimensione Formato  
tesi_ManenteDef.pdf

accesso aperto

Dimensione 18.81 MB
Formato Adobe PDF
18.81 MB Adobe PDF Visualizza/Apri

The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12608/21569