A novel method for learning policies from constrained motion

  • Authors:
  • Matthew Howard;Stefan Klanke;Michael Gienger;Christian Goerick;Sethu Vijayakumar

  • Affiliations:
  • Institute of Perception Action and Behaviour, University of Edinburgh, Scotland, UK;Institute of Perception Action and Behaviour, University of Edinburgh, Scotland, UK;Honda Research Institute Europe (GmBH), Offenbach, Germany;Honda Research Institute Europe (GmBH), Offenbach, Germany;Institute of Perception Action and Behaviour, University of Edinburgh, Scotland, UK

  • Venue:
  • ICRA'09 Proceedings of the 2009 IEEE international conference on Robotics and Automation
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Many everyday human skills can be framed in terms of performing some task subject to constraints imposed by the environment. Constraints are usually unobservable and frequently change between contexts. In this paper, we present a novel approach for learning (unconstrained) control policies from movement data, where observations come from movements under different constraints. As a key ingredient, we introduce a small but highly effective modification to the standard risk functional, allowing us to make a meaningful comparison between the estimated policy and constrained observations. We demonstrate our approach on systems of varying complexity, including kinematic data from the ASIMO humanoid robot with 27 degrees of freedom.