Home

Awesome

MoST (CVPR 2024)

This is the official implementation of "MoST: Motion Style Transformer between Diverse Action Contents (CVPR 2024)" [paper] [project]

framework

Dependencies

We tested our code on the following environment.

Install python libraries with:

pip install -r requirements.txt

Data preparation

  1. Download raw motion dataset mocap_xia.zip to ./data/ from the following link.

  2. Unzip the dataset

    cd ./data
    unzip mocap_xia.zip
    
  3. Preprocess the data

    python xia_preprocess.py
    

    Three folders will be generated.

    • preprocessed_xia : Preprocessed training data and distribution data.
    • preprocessed_xia_test: Copies of testing files.
    • preprocessed_xia_gt: Copies of training files.

Training

Sample arguments for training:

(please refer to arguments.py for detailed arguments.)

python train.py --save_path results

Demo

Sample arguments for demonstration:

(please refer to arguments.py for detailed arguments.)

python demo.py \
    --model_path pretrained/xia_pretrained.pth
    --demo_datapath data/preprocessed_xia_test
    --cnt_clip angry_13_000 --sty_clip strutting_16_000 

generated_motion.bvh will be created.

Evaluation

Arguments for calculate Content Consistency (CC):

python evaluation_CC.py \
	--model_path pretrained/xia_pretrained.pth

Arguments for calculate Style Consistency ++(SC++):

python evaluation_SCpp.py \
	--model_path pretrained/xia_pretrained.pth

Reference

Part of our code is based on GL-Transformer, Aberman et. al., and Park et. al.

Thanks to the great resources.

Citation

Please cite our work if you find it useful.

@inproceedings{kim2024most,
  title={MoST: Motion Style Transformer between Diverse Action Contents},
  author={Kim, Boeun and Kim, Jungho and Chang, Hyung Jin and Choi, Jin Young},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={1705--1714},
  year={2024}
}

Please cite the below work if you use the code, as MoST adopts the transformer architecture of GL-Transformer.

@inproceedings{kim2022global,
  title={Global-local motion transformer for unsupervised skeleton-based action learning},
  author={Kim, Boeun and Chang, Hyung Jin and Kim, Jungho and Choi, Jin Young},
  booktitle={Computer Vision--ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23--27, 2022, Proceedings, Part IV},
  pages={209--225},
  year={2022},
  organization={Springer}
}