Home

Awesome

<p align="center"> <img src="./media/psy_logo.png"> </p> <h1 align="center">D-IF: Uncertainty-aware Human Digitization via Implicit Distribution Field [<i>ICCV2023</i>]</h1>

Official PyTorch implementation for the paper:

D-IF: Uncertainty-aware Human Digitization via Implicit Distribution Field. ICCV 2023.

Xueting Yang*, Yihao Luo*, Yuliang Xiu, Wei Wang, Hao Xu, Zhaoxin Fan (*equally contributed)

<a href='https://arxiv.org/abs/2308.08857'><img src='https://img.shields.io/badge/arXiv-2308.08857-red'></a> <a href='https://yxt7979.github.io/idf/'><img src='https://img.shields.io/badge/Project-Video-Green'></a> License ↗

<i><strong>Implicit Distance Field → Implicit Distribution Field</strong>, pointwise uncertainty is modeled as the magnitude of variance in Gaussian. This simple trick enhances multiple baselines (PIFu, PaMIR, ICON) significantly.</i>

<p align="center"> <img src="./media/DIF-pipeline .png" width="100%" /> </p>

Environment

Clone the repo:

git clone https://github.com/psyai-net/D-IF_release.git
cd D-IF_release

Create conda environment:

conda env create -f environment.yaml
conda init bash
source ~/.bashrc
source activate D-IF
pip install -r requirements.txt --use-deprecated=legacy-resolver

Demo

The pretrained D-IF model has been uploaded here. You could change the resume path in ./configs/d_if.yaml.

python -m apps.infer -cfg ./configs/d_if.yaml -gpu 0 -in_dir ./examples -out_dir ./results -export_video -loop_smpl 100 -loop_cloth 200 -hps_type pixie

Training & Testing

Train dataset: Thuman2.0, for download, please follow the steps of ICON_train completely.

CUDA_VISIBLE_DEVICES=7 python -m apps.train -cfg ./configs/train/d_if.yaml

Test dataset: CAPE, for download, please follow the steps of ICON_test completely.

python -m apps.train -cfg ./configs/train/d_if.yaml -test

Citation

If you find this work useful for your research, please cite our paper:

@inproceedings{yang2023dif,
  title={{D-IF: Uncertainty-aware Human Digitization via Implicit Distribution Field}},
  author={Yang, Xueting and Luo, Yihao and Xiu, Yuliang and Wang, Wei and Xu, Hao and Fan, Zhaoxin},
  booktitle={Proceedings of the IEEE/CVF international conference on computer vision},
  year={2023}
}

Acknowledgement

Here are some great resources we benefit:

Contact

For research purpose, please contact xueting.yang99@gmail.com

For commercial licensing, please contact fanzhaoxin@psyai.net and ps-licensing@tue.mpg.de

License

This project is licensed under the Creative Commons Attribution-NonCommercial 4.0 International License. Please read the LICENSE file for more information.

Invitation

We invite you to join Psyche AI Inc to conduct cutting-edge research and business implementation together. At Psyche AI Inc, we are committed to pushing the boundaries of what's possible in the fields of artificial intelligence and computer vision, especially their applications in avatars. As a member of our team, you will have the opportunity to collaborate with talented individuals, innovate new ideas, and contribute to projects that have a real-world impact.

If you are passionate about working on the forefront of technology and making a difference, we would love to hear from you. Please visit our website at Psyche AI Inc to learn more about us and to apply for open positions. You can also contact us by fanzhaoxin@psyai.net.

Let's shape the future together!!