Abstract
Visual planning approaches have shown great success for decision making tasks with no explicit model of the state space. Learning a suitable representation and constructing a latent space where planning can be performed allows non-experts to setup and plan motions by just providing images. However, learned latent spaces are usually not semantically-interpretable, and thus it is difficult to integrate task constraints. We propose a novel framework to determine whether plans satisfy constraints given demonstrations of policies that satisfy or violate the constraints. The demonstrations are realizations of Linear Temporal Logic formulas which are employed to train Long Short-Term Memory (LSTM) networks directly in the latent space representation. We demonstrate that our architecture enables designers to easily specify, compose and integrate task constraints and achieves high performance in terms of accuracy. Furthermore, this visual planning framework enables human interaction, coping the environment changes that a human worker may involve. We show the flexibility of the method on a box pushing task in a simulated warehouse setting with different task constraints.
Original language | English |
---|---|
Title of host publication | 2021 30th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN 2021 |
Publisher | IEEE |
Pages | 131-138 |
ISBN (Electronic) | 9781665404921 |
DOIs | |
Publication status | Published - 2021 |
Externally published | Yes |
Event | 30th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN 2021 - Virtual, Vancouver, Canada Duration: 8 Aug 2021 → 12 Aug 2021 |
Conference
Conference | 30th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN 2021 |
---|---|
Country/Territory | Canada |
City | Virtual, Vancouver |
Period | 8/08/21 → 12/08/21 |