HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Diffusion Motion: Generate Text-Guided 3D Human Motion by Diffusion Model

Zhiyuan Ren Zhihong Pan Xin Zhou Le Kang

Diffusion Motion: Generate Text-Guided 3D Human Motion by Diffusion Model

Abstract

We propose a simple and novel method for generating 3D human motion from complex natural language sentences, which describe different velocity, direction and composition of all kinds of actions. Different from existing methods that use classical generative architecture, we apply the Denoising Diffusion Probabilistic Model to this task, synthesizing diverse motion results under the guidance of texts. The diffusion model converts white noise into structured 3D motion by a Markov process with a series of denoising steps and is efficiently trained by optimizing a variational lower bound. To achieve the goal of text-conditioned image synthesis, we use the classifier-free guidance strategy to fuse text embedding into the model during training. Our experiments demonstrate that our model achieves competitive results on HumanML3D test set quantitatively and can generate more visually natural and diverse examples. We also show with experiments that our model is capable of zero-shot generation of motions for unseen text guidance.

Benchmarks

BenchmarkMethodologyMetrics
motion-synthesis-on-humanml3dDiffuion Motion
Diversity: 23.692
FID: 10.21
R Precision Top3: 0.735

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Diffusion Motion: Generate Text-Guided 3D Human Motion by Diffusion Model | Papers | HyperAI