Methods for Learning Control Policies from Variable-constraint Demonstrations

M. Howard, S. Klanke, M. Gienger, C. Goerick, S. Vijayakumar

Research output: Chapter in Book/Report/Conference proceedingOther chapter contribution

8 Citations (Scopus)

Abstract

Many everyday human skills can be framed in terms of performing some task subject to constraints imposed by the task or the environment. Constraints are usually not observable and frequently change between contexts. In this chapter, we explore the problem of learning control policies from data containing variable, dynamic and non-linear constraints on motion. We discuss how an effective approach for doing this is to learn the unconstrained policy in a way that is consistent with the constraints. We then go on to discuss several recent algorithms for extracting policies from movement data, where observations are recorded under variable, unknown constraints. We review a number of experiments testing the performance of these algorithms and demonstrating how the resultant policy models generalise over constraints allowing prediction of behaviour under unseen settings where new constraints apply.
Original languageEnglish
Title of host publicationFrom Motor to Interaction Learning in Robots
EditorsOlivier Sigaud, Jan Peters
Place of PublicationBerlin
PublisherSpringer Finance
Pages253-291
Number of pages39
VolumeN/A
EditionN/A
ISBN (Print)9783642051807
DOIs
Publication statusPublished - 2010

Publication series

NameStudies in Computational Science
PublisherSpringer Berlin Heidelberg
Volume264
ISSN (Print)1860-949X

Fingerprint

Dive into the research topics of 'Methods for Learning Control Policies from Variable-constraint Demonstrations'. Together they form a unique fingerprint.

Cite this