Model-based and model-free perspectives are two well established paradigms in RL. in this thesis a mixed approach is proposed, in which the interactions with the real system are carried out in both ways: a rough model is retrieved in order to play the role of a regularizer, while the punctual estimation over specific values of the policy parameter is placing reliable punctual estimates that should be fitted by the reconstructed function.

System Identification meets Reinforcement Learning: probabilistic dynamics for regularization

Zanini, Francesco
2019/2020

Abstract

Model-based and model-free perspectives are two well established paradigms in RL. in this thesis a mixed approach is proposed, in which the interactions with the real system are carried out in both ways: a rough model is retrieved in order to play the role of a regularizer, while the punctual estimation over specific values of the policy parameter is placing reliable punctual estimates that should be fitted by the reconstructed function.
2019-09-10
learning, regularization
File in questo prodotto:
File Dimensione Formato  
Tesi_Zanini.pdf

Open Access dal 11/09/2022

Dimensione 1.61 MB
Formato Adobe PDF
1.61 MB Adobe PDF Visualizza/Apri

The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12608/28897