Learning Human Preferences for Physical Human-Robot Cooperation

L.F. van der Spaa

Research output: ThesisDissertation (TU Delft)

18 Downloads (Pure)

Abstract

Physical human-robot cooperation (pHRC) has the potential to combine human and robot strengths in a team that can achieve more than a human and a robot working on the task separately. However, how much of the potential can be realized depends on the quality of cooperation, in which awarenes of the partner’s intention and preferences plays an important role. Preferences tend to be highly personal, and additionally depend on the cooperation partner and the cooperation itself. They can be hard to define in terms a robot would understand, and may change over time. This thesis focuses on learning ‘useful models’ from observed behavior, to let our robot adapt its behavior to better match its human partner’s preferences, and thus improve the cooperation.
The aim is to capture personalized approximate models of human preferences –how a person likes to do something– from very few interactive observations, providing only small amounts of imprecise data, such that the robot can use the model to improve each user’s comfort. First, we learn a model to predict and optimize the human ergonomics in a pHRC task, such that our robot can ropose a plan, for both the human and itself, to solve the task in a way that is more ergonomic for its human partner. However, people do not necessarily prefer to act ergonomically, nor do we want to impose on them what a robot thinks best. Therefore, next, we apply inverse reinforcement learning (IRL), to capture less restrictive preference models: 1) path and velocity preferences for motion planning, and 2) on a higher level of abstraction, which (grasp or motion) action to initiate for proactive physical support. For learning to take the correct action in cooperation, we developed the disagreement-aware variable impedance (DAVI) controller to smoothly transition between providing active guidance and allowing the human to demonstrate alternative behavior.....
Original languageEnglish
QualificationDoctor of Philosophy
Awarding Institution
  • Delft University of Technology
Supervisors/Advisors
  • Kober, J., Supervisor
  • Babuska, R., Supervisor
Thesis sponsors
Award date1 Feb 2024
Print ISBNs978-94-6483-764-3
Electronic ISBNs978-94-6483-779-7
DOIs
Publication statusPublished - 2024

Bibliographical note

Dr. M. Gienger contributed significantly to the realization of the dissertation.

Keywords

  • Physical Human-Robot Interaction
  • Human-Robot Collaboration
  • human preferences
  • human-centered planning
  • Inverse Reinforcement Learning

Fingerprint

Dive into the research topics of 'Learning Human Preferences for Physical Human-Robot Cooperation'. Together they form a unique fingerprint.

Cite this