Abstract
We present Multitask Soft Option Learning (MSOL), a hierarchical multitask framework based on Planning as Inference. MSOL extends the concept of options, using separate variational posteriors for each task, regularized by a shared prior. This “soft” version of options avoids several instabilities during training in a multitask setting, and provides a natural way to learn both intra-option policies and their terminations. Furthermore, it allows fine-tuning of options for new tasks without forgetting their learned policies, leading to faster training without reducing the expressiveness of the hierarchical policy. We demonstrate empirically that MSOL significantly outperforms both hierarchical and flat transfer-learning baselines.
Original language | English |
---|---|
Title of host publication | Proceedings of the 36th Conference on Uncertainty in Artificial Intelligence (UAI) |
Pages | 969-978 |
Number of pages | 10 |
Volume | 124 |
Publication status | Published - 2020 |
Event | 36th Conference on Uncertainty in Artificial Intelligence - Virtual/online event Duration: 4 Aug 2020 → 6 Aug 2020 Conference number: 36 |
Publication series
Name | Proceedings of Machine Learning Research |
---|
Conference
Conference | 36th Conference on Uncertainty in Artificial Intelligence |
---|---|
Abbreviated title | UAI 2020 |
Period | 4/08/20 → 6/08/20 |