Skip to main content Skip to main navigation


Stylistic Locomotion Modeling and Synthesis using Variational Generative Models

Han Du; Erik Herrmann; Janis Sprenger; Klaus Fischer; Philipp Slusallek
In: Hubert P. H. Shum; Edmond S. L. Ho; Marie-Paule Cani; Tiberiu Popa; Daniel Holden; He Wang (Hrsg.). Motion, Interaction and Games 2019. ACM SIGGRAPH Conference on Motion in Games (MIG-2019), October 28-30, Newcastle upon Tyne, United Kingdom, ISBN 978-1-4503-6994-7, ACM, New York, NY, USA, 10/2019.


We propose a novel approach to create generative models for distinctive styles of locomotion for humanoid characters. Our approach only requires a single or a few style examples and a neutral motion database. We are inspired by the observation that human styles can be easily distinguished from a few examples. However, learning a generative model for natural human motions which can display huge amounts of variations and randomness would require a lot of training data. Furthermore, it would require considerable efforts to create such a large motion database for each style. One solution for that is motion style transfer, which provides the possibility of converting the content of the motion from one style to the other. Typically style transfer focuses on transferring the content motion to target style explicitly. We propose a variational generative model to combine the large variation in neutral motion database and style information from a limited number of examples. We formulate the style motion modeling as a conditional distribution learning problem and style transfer is implicitly applied during the model learning process. A conditional variational autoencoder (CVAE) is applied to learn the distribution and stylistic examples are used as constraints. We demonstrate that our approach can generate any number of natural-looking, various human motions with a similar style to the target.


Weitere Links