Abstract
Recent years have seen a surge of algorithms and architectures for deep Re-
inforcement Learning (RL), many of which have shown remarkable success for
various problems. Yet, little work has attempted to relate the performance of
these algorithms and architectures to what the resulting deep RL agents actu-
ally learn, and whether this corresponds to what they should ideally learn. Such
a comparison may allow for both an improved understanding of why certain
algorithms or network architectures perform better than others and the devel-
opment of methods that specically address discrepancies between what is and
what should be learned.
inforcement Learning (RL), many of which have shown remarkable success for
various problems. Yet, little work has attempted to relate the performance of
these algorithms and architectures to what the resulting deep RL agents actu-
ally learn, and whether this corresponds to what they should ideally learn. Such
a comparison may allow for both an improved understanding of why certain
algorithms or network architectures perform better than others and the devel-
opment of methods that specically address discrepancies between what is and
what should be learned.
Original language | English |
---|---|
Number of pages | 3 |
Publication status | Published - 2020 |
Event | BNAIC/BENELEARN 2020 - Leiden, Netherlands Duration: 19 Nov 2020 → 20 Nov 2020 |
Conference
Conference | BNAIC/BENELEARN 2020 |
---|---|
Country/Territory | Netherlands |
City | Leiden |
Period | 19/11/20 → 20/11/20 |
Keywords
- Deep Reinforcement Learning
- Representation Learning
- Bisimulation Metrics
- Markovianity