Search for a command to run...
Many control problems take place in continuous state-action spaces, e.g., as in manipulator robotics, where the control objective is of-ten de¯ned as ¯nding a desired trajectory that reaches a particular goal state. While reinforcement learning o®ers a theoretical frame-work to learn such control policies from scratch, its applicability to higher dimensional continuous state-action spaces remains rather limited to date. Instead of learning from scratch, in this paper we suggest to learn a desired complex control policy by transforming an existing simple canonical control policy. For this purpose, we represent canonical policies in terms of di®erential equations with well-de¯ned attractor properties. By nonlinearly transforming the canonical attractor dynamics using techniques from nonparametric regression, almost arbitrary new nonlinear policies can be gener-ated without losing the stability properties of the canonical sys-tem. We demonstrate our techniques in the context of learning a set of movement skills for a humanoid robot from demonstrations of a human teacher. Policies are acquired rapidly, and, due to the properties of well formulated di®erential equations, can be re-used and modi¯ed on-line under dynamic changes of the environment. The linear parameterization of nonparametric regression moreover lends itself to recognize and classify previously learned movement skills. Evaluations in simulations and on an actual 30 degree-of-freedom humanoid robot exemplify the feasibility and robustness of our approach. 1