In standard reinforcement learning (RL) an agent optimizes its policy by exploring the environment via trial and error which can lead to potentially dangerous situations. In this work, we investigate an approach for enabling an RL agent to learn about dangerous states or constraints in the environment from stop-feedback, i.e., feedback preventing the agent from taking any further, potentially dangerous, actions. In many practical scenarios, we would assume that the stop-feedback is provided by human supervisors giving feedback on the RL agent's behavior while carrying out complex tasks in real-world scenarios, which, without this feedback, would require the design of sophisticated and overly complicated reward functions. To enable the RL agent to learn from the supervisor’s feedback, we develop a simple probabilistic model for approximating how the supervisor’s feedback is generated and consider a Bayesian approach for inferring dangerous states. We evaluate our approach in experiments using the OpenAI Safety Gym environment and demonstrate that our agent can effectively follow the imposed safety requirements. Furthermore, we gathered feedback from 100 human volunteers to validate our human-like feedback model and to obtain insights into the human provision of stop-feedback.
In standard reinforcement learning (RL) an agent optimizes its policy by exploring the environment via trial and error which can lead to potentially dangerous situations. In this work, we investigate an approach for enabling an RL agent to learn about dangerous states or constraints in the environment from stop-feedback, i.e., feedback preventing the agent from taking any further, potentially dangerous, actions. In many practical scenarios, we would assume that the stop-feedback is provided by human supervisors giving feedback on the RL agent's behavior while carrying out complex tasks in real-world scenarios, which, without this feedback, would require the design of sophisticated and overly complicated reward functions. To enable the RL agent to learn from the supervisor’s feedback, we develop a simple probabilistic model for approximating how the supervisor’s feedback is generated and consider a Bayesian approach for inferring dangerous states. We evaluate our approach in experiments using the OpenAI Safety Gym environment and demonstrate that our agent can effectively follow the imposed safety requirements. Furthermore, we gathered feedback from 100 human volunteers to validate our human-like feedback model and to obtain insights into the human provision of stop-feedback.
Learning constraints from human stop-feedbacks in Reinforcement Learning
POLETTI, SILVIA
2021/2022
Abstract
In standard reinforcement learning (RL) an agent optimizes its policy by exploring the environment via trial and error which can lead to potentially dangerous situations. In this work, we investigate an approach for enabling an RL agent to learn about dangerous states or constraints in the environment from stop-feedback, i.e., feedback preventing the agent from taking any further, potentially dangerous, actions. In many practical scenarios, we would assume that the stop-feedback is provided by human supervisors giving feedback on the RL agent's behavior while carrying out complex tasks in real-world scenarios, which, without this feedback, would require the design of sophisticated and overly complicated reward functions. To enable the RL agent to learn from the supervisor’s feedback, we develop a simple probabilistic model for approximating how the supervisor’s feedback is generated and consider a Bayesian approach for inferring dangerous states. We evaluate our approach in experiments using the OpenAI Safety Gym environment and demonstrate that our agent can effectively follow the imposed safety requirements. Furthermore, we gathered feedback from 100 human volunteers to validate our human-like feedback model and to obtain insights into the human provision of stop-feedback.File | Dimensione | Formato | |
---|---|---|---|
dissertation_pdfa.pdf
accesso riservato
Dimensione
3.28 MB
Formato
Adobe PDF
|
3.28 MB | Adobe PDF |
The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License
https://hdl.handle.net/20.500.12608/42068