Home

Awesome

StyledMotionSynthesis

Code for the GRAPP 2023 paper "Unifying Human Motion Synthesis and Style Transfer with Denoising Diffusion Probabilistic Models"

Environment Setup

Please create an appropriate environment using conda:

conda env create -f environment.yml

conda activate motiondiffusion

Download Data

Xia dataset can be downloaded from https://github.com/DeepMotionEditing/deep-motion-editing and extracted into ./data.

Run data_proc/gen_dataset.sh to generate the dataset.

Citation

@inproceedings{chang2022unifying,
    title={Unifying Human Motion Synthesis and Style Transfer with Denoising Diffusion Probabilistic Models},
    author={Chang, Ziyi and Findlay, Edmund J. C. and Zhang, Haozheng and Shum, Hubert P. H.},
    booktitle={Proceedings of the 2023 International Conference on Computer Graphics Theory and Applications},
    year={2022}
}