Home

Awesome

HybrIK: Hybrid Analytical-Neural Inverse Kinematics for Body Mesh Recovery

<div align="center"> <img src="assets/hybrik_dance1.gif" height="160"> <img src="assets/hybrik_dance2.gif" height="160"> <img src="assets/hybrik_dance3.gif" height="160"> </br> <img src="assets/hybrikx_dance1.gif" height="160"> <img src="assets/hybrikx_basketball.gif" height="160"> </div> <div align="center"> <a href="https://pytorch.org/get-started/locally/"><img alt="PyTorch" src="https://img.shields.io/badge/PyTorch-ee4c2c?logo=pytorch&logoColor=white"></a> <a href='https://colab.research.google.com/drive/1n41l7I2NxWseuruVQEU8he2XqzSXhu2f?usp=sharing' style='padding-left: 0.5rem;'><img src='https://colab.research.google.com/assets/colab-badge.svg' alt='Google Colab'></a> <a href='https://paperswithcode.com/sota/3d-human-pose-estimation-on-agora?p=hybrik-x-hybrid-analytical-neural-inverse' style='padding-left: 0.5rem;'><img src='https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/hybrik-x-hybrid-analytical-neural-inverse/3d-human-pose-estimation-on-agora'> </a> </br> <a href="https://arxiv.org/abs/2011.14672"> <img src='https://img.shields.io/badge/Paper-HybrIK-green?style=for-the-badge&logo=adobeacrobatreader&logoWidth=20&logoColor=white&labelColor=66cc00&color=94DD15' alt='Paper PDF'> </a> <a href="https://arxiv.org/abs/2304.05690"> <img src='https://img.shields.io/badge/Paper-HybrIK--X-green?style=for-the-badge&logo=adobeacrobatreader&logoWidth=20&logoColor=white&labelColor=3CB371&color=40E0D0' alt='Paper PDF'> </a> <a href="https://youtu.be/xbdNFw2eZ4o"> <img alt="youtube views" src="https://img.shields.io/badge/Demo-HybrIK-red?logo=youtube&logoColor=white&labelColor=D35400&style=for-the-badge"/> </a> <a href="https://youtu.be/VzCXKdQSZ7c"> <img alt="youtube views" src="https://img.shields.io/badge/Demo-HybrIK--X-red?logo=youtube&logoColor=white&labelColor=D35400&style=for-the-badge"/> </a> </br> </div>

This repo contains the code of our papers:

HybrIK: A Hybrid Analytical-Neural Inverse Kinematics Solution for 3D Human Pose and Shape Estimation, In CVPR 2021

HybrIK-X: Hybrid Analytical-Neural Inverse Kinematics for Whole-body Mesh Recovery, ArXiv 2023

<!-- ![hybrik](assets/hybrik.png) --> <!-- <div align="center"> <img src="assets/decompose.gif", width="600" alt><br> Twist-and-Swing Decomposition </div> -->

News :triangular_flag_on_post:

[2023/06/02] Demo code for whole-body HybrIK-X is released.

[2022/12/03] HybrIK for Blender add-on is now available for download. The output of HybrIK can be imported to Blender and saved as fbx.

[2022/08/16] Pretrained model with HRNet-W48 backbone is available.

[2022/07/31] Training code with predicted camera is released.

[2022/07/25] HybrIK is now supported in Alphapose! Multi-person demo with pose-tracking is available.

[2022/04/27] <a href='https://colab.research.google.com/drive/1n41l7I2NxWseuruVQEU8he2XqzSXhu2f?usp=sharing' style='padding-left: 0.5rem;'><img src='https://colab.research.google.com/assets/colab-badge.svg' alt='Google Colab'></a> is ready to use.

[2022/04/26] Achieve SOTA results by adding the 3DPW dataset for training.

[2022/04/25] The demo code is released!

Key idea: Inverse Kinematics

HybrIK and HybrIK-X are based on a hybrid inverse kinematics (IK) to convert accurate 3D keypoints to parametric body meshes.

<div align="center"> <img src="assets/ik.gif", width="600" alt><br> Twist-and-Swing Decomposition </div>

Installation instructions

# 1. Create a conda virtual environment.
conda create -n hybrik python=3.8 -y
conda activate hybrik

# 2. Install PyTorch
conda install pytorch==1.9.1 torchvision==0.10.1 -c pytorch

# 3. Install PyTorch3D (Optional, only for visualization)
conda install -c fvcore -c iopath -c conda-forge fvcore iopath
conda install -c bottler nvidiacub
pip install git+ssh://git@github.com/facebookresearch/pytorch3d.git@stable

# 4. Pull our code
git clone https://github.com/Jeff-sjtu/HybrIK.git
cd HybrIK

# 5. Install
pip install pycocotools
python setup.py develop  # or "pip install -e ."

Download necessary model files from [Google Drive | Baidu (code: 2u3c) ] and un-zip them in the ${ROOT} directory.

MODEL ZOO

HybrIK (SMPL)

BackboneTraining DataPA-MPJPE (3DPW)MPJPE (3DPW)PA-MPJPE (Human3.6M)MPJPE (Human3.6M)DownloadConfig
ResNet-34w/ 3DPW44.572.433.855.5modelcfg
HRNet-W48w/o 3DPW48.688.029.550.4modelcfg
HRNet-W48w/ 3DPW41.871.329.847.1modelcfg

HybrIK-X (SMPL-X)

BackboneMVE (AGORA Test)MPJPE (AGORA Test)DownloadConfig
HRNet-W48134.1127.5modelcfg
HRNet-W48 + RLE112.1107.6modelcfg

Demo

First make sure you download the pretrained model (with predicted camera) and place it in the ${ROOT}/pretrained_models directory, i.e., ./pretrained_models/hybrik_hrnet.pth and ./pretrained_models/hybrikx_rle_hrnet.pth.

SMPL

python scripts/demo_video.py --video-name examples/dance.mp4 --out-dir res_dance --save-pk --save-img

The saved results in ./res_dance/res.pk can be imported to Blender with our add-on.

python scripts/demo_image.py --img-dir examples --out-dir res

SMPL-X

python scripts/demo_video_x.py --video-name examples/dance.mp4 --out-dir res_dance --save-pk --save-img

Fetch data

Download Human3.6M, MPI-INF-3DHP, 3DPW and MSCOCO datasets. You need to follow directory structure of the data as below. Thanks to the great job done by Moon et al., we use the Human3.6M images provided in PoseNet.

|-- data
`-- |-- h36m
    `-- |-- annotations
        `-- images
`-- |-- pw3d
    `-- |-- json
        `-- imageFiles
`-- |-- 3dhp
    `-- |-- annotation_mpi_inf_3dhp_train.json
        |-- annotation_mpi_inf_3dhp_test.json
        |-- mpi_inf_3dhp_train_set
        `-- mpi_inf_3dhp_test_set
`-- |-- coco
    `-- |-- annotations
        |   |-- person_keypoints_train2017.json
        |   `-- person_keypoints_val2017.json
        |-- train2017
        `-- val2017

Train from scratch

./scripts/train_smpl_cam.sh test_3dpw configs/256x192_adam_lr1e-3-res34_smpl_3d_cam_2x_mix_w_pw3d.yaml

Evaluation

Download the pretrained model (ResNet-34 or HRNet-W48).

./scripts/validate_smpl_cam.sh ./configs/256x192_adam_lr1e-3-hrw48_cam_2x_w_pw3d_3dhp.yaml ./pretrained_hrnet.pth

Citing

If our code helps your research, please consider citing the following paper:

@inproceedings{li2021hybrik,
    title={Hybrik: A hybrid analytical-neural inverse kinematics solution for 3d human pose and shape estimation},
    author={Li, Jiefeng and Xu, Chao and Chen, Zhicun and Bian, Siyuan and Yang, Lixin and Lu, Cewu},
    booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
    pages={3383--3393},
    year={2021}
}

@article{li2023hybrik,
    title={HybrIK-X: Hybrid Analytical-Neural Inverse Kinematics for Whole-body Mesh Recovery},
    author={Li, Jiefeng and Bian, Siyuan and Xu, Chao and Chen, Zhicun and Yang, Lixin and Lu, Cewu},
    journal={arXiv preprint arXiv:2304.05690},
    year={2023}
}