Risk Aversion and Guided Exploration in Safety-Constrained Reinforcement Learning

Research output: ThesisDissertation (TU Delft)

85 Downloads (Pure)

Abstract

In traditional reinforcement learning (RL) problems, agents can explore environments to learn optimal policies through trials and errors that are sometimes unsafe. However, unsafe interactions with environments are unacceptable in many safety-critical problems, for instance in robot navigation tasks. Even though RL agents can be trained in simulators, there are many real-world problems without simulators of sufficient fidelity. Constructing safe exploration algorithms for dangerous environments is challenging because we have to optimize policies under the premise of safety. In general, safety is still an open problem that hinders the wider application of RL.
Original languageEnglish
QualificationDoctor of Philosophy
Awarding Institution
  • Delft University of Technology
Supervisors/Advisors
  • Spaan, M.T.J., Supervisor
  • Tindemans, Simon H., Advisor
Award date23 Jun 2023
Electronic ISBNs978-94-6384-458-1
DOIs
Publication statusPublished - 2023

Keywords

  • Reinforcement Leaning (RL)
  • constrained optimization
  • quantile regression
  • taskagnostic exploration

Fingerprint

Dive into the research topics of 'Risk Aversion and Guided Exploration in Safety-Constrained Reinforcement Learning'. Together they form a unique fingerprint.

Cite this