Learning Reduced-Order Feedback Policies for Motion Skills

Loading...
Thumbnail Image
Date
2015
Journal Title
Journal ISSN
Volume Title
Publisher
ACM Siggraph
Abstract
We introduce a method for learning low-dimensional linear feedback strategies for the control of physics-based animated characters around a given reference trajectory. This allows for learned low-dimensional state abstractions and action abstractions, thereby reducing the need to rely on manually designed abstractions such as the center-of-mass state or foot-placement actions. Once learned, the compact feedback structure allow simulated characters to respond to changes in the environment and changes in goals. The approach is based on policy search in the space of reduced-order linear output feedback matrices. We show that these can be used to replace or further reduce manually-designed state and action abstractions. The approach is sufficiently general to allow for the development of unconventional feedback loops, such as feedback based on ground reaction forces. Results are demonstrated for a mix of 2D and 3D systems, including tilting-platform balancing, walking, running, rolling, targeted kicks, and several types of ballhitting tasks.
Description

        
@inproceedings{
10.1145:2786784.2786802
, booktitle = {
ACM/ Eurographics Symposium on Computer Animation
}, editor = {
Florence Bertails-Descoubes and Stelian Coros and Shinjiro Sueda
}, title = {{
Learning Reduced-Order Feedback Policies for Motion Skills
}}, author = {
Ding, Kai
and
Liu, Libin
and
Panne, Michiel van de
and
Yin, KangKang
}, year = {
2015
}, publisher = {
ACM Siggraph
}, ISBN = {
978-1-4503-3496-9
}, DOI = {
10.1145/2786784.2786802
} }
Citation