Home

Awesome

GFGE

Teaser image

This is a repository with training and inference code for the paper ["Audio-Driven Stylized Gesture Generation with Flow-Based Model"].

Requirements

Getting started

Datasets

In this work, we conducted our experiments on two datasets: TED Dataset and Trinity Dataset.

Feature Extractors

Model Checkpoints

We provide several pretrained model checkpoints. Download and extract these ZIP files into ./results.

Usage

First, please make sure that all requirements are satisfied and all required files are downloaded (see above steps).

Train

# train on ted dataset
python scripts/train.py hparams/preferred/locomotion.json locomotion

# train on trinity dataset
python scripts/train.py hparams/preferred/trinity.json trinity

Sample

# sample on ted dataset
python scripts/test_locomotion_sample.py

# sample on trinity dataset
python scripts/test_trinity_sample.py

Evaluate

python scripts/cal_metrics.py

Latent Space Visualization

python scripts/vis_latent_space.py

Style Transfer

python scripts/style_transfer.py

Results

TED Trinity

Acknowledgement

Note that the training and testing code of this repo is heavily rely on MoGlow and GTC. We thank the authors for their great job!