Back to Search Start Over

GANimator:Neural Motion Synthesis from a Single Sequence

Authors :
Li, Peizhuo
Aberman, Kfir
Zhang, Zihan
Hanocka, Rana
Sorkine-Hornung, Olga
Source :
ACM Transactions on Graphics, 41 (4)
Publication Year :
2022
Publisher :
Association for Computing Machinery, 2022.

Abstract

We present GANimator, a generative model that learns to synthesize novel motions from a single, short motion sequence. GANimator generates motions that resemble the core elements of the original motion, while simultaneously synthesizing novel and diverse movements. Existing data-driven techniques for motion synthesis require a large motion dataset which contains the desired and specific skeletal structure. By contrast, GANimator only requires training on a single motion sequence, enabling novel motion synthesis for a variety of skeletal structures e.g., bipeds, quadropeds, hexapeds, and more. Our framework contains a series of generative and adversarial neural networks, each responsible for generating motions in a specific frame rate. The framework progressively learns to synthesize motion from random noise, enabling hierarchical control over the generated motion content across varying levels of detail. We show a number of applications, including crowd simulation, key-frame editing, style transfer, and interactive control, which all learn from a single input sequence. Code and data for this paper are at https://peizhuoli.github.io/ganimator.<br />ACM Transactions on Graphics, 41 (4)<br />ISSN:0730-0301<br />ISSN:1557-7368

Details

Language :
English
ISSN :
07300301 and 15577368
Database :
OpenAIRE
Journal :
ACM Transactions on Graphics, 41 (4)
Accession number :
edsair.doi.dedup.....17a0c7486bc24a6e7ad9780dd4c54be5