In this thesis I analyse Deep Learning techniques for model-based Reinforcement Learning with sensory data comprising of images and scalar variables. The methods are based on world models: nonlinear stochastic state-space models, implemented as artificial neural networks, that try to capture the spatial and temporal aspects of the agent/environment interaction. World models can be used to generate dream environments to learn a controller using Reinforcement Learning algorithms without the need of interacting with the real environments. Experiments are conducted in two OpenAI Gym environments and in the open-source CARLA simulator. The latter is used to investigate how world models might be applied in the context of self-driving vehicles, including approaches for action planning, for predicting various potential futures based on actions, and for choosing control inputs.

In this thesis I analyse Deep Learning techniques for model-based Reinforcement Learning with sensory data comprising of images and scalar variables. The methods are based on world models: nonlinear stochastic state-space models, implemented as artificial neural networks, that try to capture the spatial and temporal aspects of the agent/environment interaction. World models can be used to generate dream environments to learn a controller using Reinforcement Learning algorithms without the need of interacting with the real environments. Experiments are conducted in two OpenAI Gym environments and in the open-source CARLA simulator. The latter is used to investigate how world models might be applied in the context of self-driving vehicles, including approaches for action planning, for predicting various potential futures based on actions, and for choosing control inputs.

Dream to drive: Deep Learning for planning, prediction and control of self-driving vehicles

GUIDOLIN, SIMONE
2021/2022

Abstract

In this thesis I analyse Deep Learning techniques for model-based Reinforcement Learning with sensory data comprising of images and scalar variables. The methods are based on world models: nonlinear stochastic state-space models, implemented as artificial neural networks, that try to capture the spatial and temporal aspects of the agent/environment interaction. World models can be used to generate dream environments to learn a controller using Reinforcement Learning algorithms without the need of interacting with the real environments. Experiments are conducted in two OpenAI Gym environments and in the open-source CARLA simulator. The latter is used to investigate how world models might be applied in the context of self-driving vehicles, including approaches for action planning, for predicting various potential futures based on actions, and for choosing control inputs.
2021
Dream to drive: Deep Learning for planning, prediction and control of self-driving vehicles
In this thesis I analyse Deep Learning techniques for model-based Reinforcement Learning with sensory data comprising of images and scalar variables. The methods are based on world models: nonlinear stochastic state-space models, implemented as artificial neural networks, that try to capture the spatial and temporal aspects of the agent/environment interaction. World models can be used to generate dream environments to learn a controller using Reinforcement Learning algorithms without the need of interacting with the real environments. Experiments are conducted in two OpenAI Gym environments and in the open-source CARLA simulator. The latter is used to investigate how world models might be applied in the context of self-driving vehicles, including approaches for action planning, for predicting various potential futures based on actions, and for choosing control inputs.
File in questo prodotto:
File Dimensione Formato  
Guidolin_Simone.pdf

accesso riservato

Dimensione 17.43 MB
Formato Adobe PDF
17.43 MB Adobe PDF

The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12608/36684