TY - JOUR
T1 - Visually-guided motion planning for autonomous driving from interactive demonstrations
AU - Pérez-Dattari, Rodrigo
AU - Brito, Bruno
AU - de Groot, Oscar
AU - Kober, Jens
AU - Alonso-Mora, Javier
PY - 2022
Y1 - 2022
N2 - The successful integration of autonomous robots in real-world environments strongly depends on their ability to reason from context and take socially acceptable actions. Current autonomous navigation systems mainly rely on geometric information and hard-coded rules to induce safe and socially compliant behaviors. Yet, in unstructured urban scenarios these approaches can become costly and suboptimal. In this paper, we introduce a motion planning framework consisting of two components: a data-driven policy that uses visual inputs and human feedback to generate socially compliant driving behaviors (encoded by high-level decision variables), and a local trajectory optimization method that executes these behaviors (ensuring safety). In particular, we employ Interactive Imitation Learning to jointly train the policy with the local planner, a Model Predictive Controller (MPC), which results in safe and human-like driving behaviors. Our approach is validated in realistic simulated urban scenarios. Qualitative results show the similarity of the learned behaviors with human driving. Furthermore, navigation performance is substantially improved in terms of safety, i.e., number of collisions, as compared to prior trajectory optimization frameworks, and in terms of data-efficiency as compared to prior learning-based frameworks, broadening the operational domain of MPC to more realistic autonomous driving scenarios.
AB - The successful integration of autonomous robots in real-world environments strongly depends on their ability to reason from context and take socially acceptable actions. Current autonomous navigation systems mainly rely on geometric information and hard-coded rules to induce safe and socially compliant behaviors. Yet, in unstructured urban scenarios these approaches can become costly and suboptimal. In this paper, we introduce a motion planning framework consisting of two components: a data-driven policy that uses visual inputs and human feedback to generate socially compliant driving behaviors (encoded by high-level decision variables), and a local trajectory optimization method that executes these behaviors (ensuring safety). In particular, we employ Interactive Imitation Learning to jointly train the policy with the local planner, a Model Predictive Controller (MPC), which results in safe and human-like driving behaviors. Our approach is validated in realistic simulated urban scenarios. Qualitative results show the similarity of the learned behaviors with human driving. Furthermore, navigation performance is substantially improved in terms of safety, i.e., number of collisions, as compared to prior trajectory optimization frameworks, and in terms of data-efficiency as compared to prior learning-based frameworks, broadening the operational domain of MPC to more realistic autonomous driving scenarios.
KW - Autonomous driving
KW - Deep learning
KW - Human in the loop
KW - Interactive Imitation Learning
KW - Model Predictive Control
KW - Motion planning
UR - http://www.scopus.com/inward/record.url?scp=85138443938&partnerID=8YFLogxK
U2 - 10.1016/j.engappai.2022.105277
DO - 10.1016/j.engappai.2022.105277
M3 - Article
AN - SCOPUS:85138443938
VL - 116
JO - Engineering Applications of Artificial Intelligence
JF - Engineering Applications of Artificial Intelligence
SN - 0952-1976
M1 - 105277
ER -