Awesome
Contents
- BiGraphGAN
- Installation
- Dataset Preparation
- Generating Images Using Pretrained Model
- Train and Test New Models
- Download Images Produced by the Authors
- Evaluation
- Acknowledgments
- Related Projects
- Citation
- Contributions
- Collaborations
BiGraphGAN
| Project | Paper | <br> Bipartite Graph Reasoning GANs for Person Image Generation <br> Hao Tang<sup>12</sup>, Song Bai<sup>2</sup>, Philip H.S. Torr<sup>2</sup>, Nicu Sebe<sup>13</sup>. <br> <sup>1</sup>University of Trento, Italy, <sup>2</sup>University of Oxford, UK, <sup>3</sup>Huawei Research Ireland, Ireland.<br> In BMVC 2020 Oral. <br> The repository offers the official implementation of our paper in PyTorch.
In the meantime, check out our related ACM MM 2019 paper Cycle In Cycle Generative Adversarial Networks for Keypoint-Guided Image Generation, ECCV 2020 paper XingGAN for Person Image Generation, ICCV 2021 paper Intrinsic-Extrinsic Preserved GANs for Unsupervised 3D Pose Transfer, and ICIAP 2021 paper Graph-based Generative Face Anonymisation with Pose Preservation!
Motivation
<img src='./imgs/motivation.jpg' width=1200>Framework
<img src='./imgs/method.jpg' width=1200>Comparison Results
<img src='./imgs/market_results.jpg' width=1200> <br> <img src='./imgs/fashion_results.jpg' width=1200>License
<a rel="license" href="http://creativecommons.org/licenses/by-nc-sa/4.0/"><img alt="Creative Commons License" style="border-width:0" src="https://i.creativecommons.org/l/by-nc-sa/4.0/88x31.png" /></a><br /> Copyright (C) 2020 University of Trento, Italy.
All rights reserved. Licensed under the CC BY-NC-SA 4.0 (Attribution-NonCommercial-ShareAlike 4.0 International)
The code is released for academic research use only. For commercial use, please contact bjdxtanghao@gmail.com.
Installation
Clone this repo.
git clone https://github.com/Ha0Tang/BiGraphGAN
cd BiGraphGAN/
This code requires PyTorch 1.0.0 and python 3.6.9+. Please install the following dependencies:
- pytorch 1.0.0
- torchvision
- numpy
- scipy
- scikit-image
- pillow
- pandas
- tqdm
- dominate
To reproduce the results reported in the paper, you need to run experiments on NVIDIA DGX1 with 4 32GB V100 GPUs for DeepFashion, and 1 32GB V100 GPU for Market-1501.
Dataset Preparation
Please follow SelectionGAN to directly download both Market-1501 and DeepFashion datasets.
This repository uses the same dataset format as SelectionGAN and XingGAN. so you can use the same data for all these methods.
Generating Images Using Pretrained Model
Market-1501
cd scripts/
sh download_bigraphgan_model.sh market
cd ..
cd market_1501/
Then,
- Change several parameters in
test_market_pretrained.sh
. - Run
sh test_market_pretrained.sh
for testing.
DeepFashion
cd scripts/
sh download_bigraphgan_model.sh deepfashion
cd ..
cd deepfashion/
Then,
- Change several parameters in
test_deepfashion_pretrained.sh
. - Run
sh test_deepfashion_pretrained.sh
for testing.
Train and Test New Models
Market-1501
- Go to the market_1501 folder.
- Change several parameters in
train_market.sh
. - Run
sh train_market.sh
for training. - Change several parameters in
test_market.sh
. - Run
sh test_market.sh
for testing.
DeepFashion
- Go to the deepfashion folder.
- Change several parameters in
train_deepfashion.sh
. - Run
sh train_deepfashion.sh
for training. - Change several parameters in
test_deepfashion.sh
. - Run
sh test_deepfashion.sh
for testing.
Evaluation
We adopt SSIM, mask-SSIM, IS, mask-IS, and PCKh for evaluation of Market-1501. SSIM, IS, PCKh for DeepFashion. Please refer to Pose-Transfer for more details.
Acknowledgments
This source code is inspired by both Pose-Transfer, and SelectionGAN.
Related Projects
XingGAN | GestureGAN | C2GAN | SelectionGAN | Guided-I2I-Translation-Papers
Citation
If you use this code for your research, please cite our papers.
BiGraphGAN
@article{tang2022bipartite,
title={Bipartite Graph Reasoning GANs for Person Pose and Facial Image Synthesis},
author={Tang, Hao and Shao, Ling and Torr, Philip HS and Sebe, Nicu},
journal={International Journal of Computer Vision (IJCV)},
year={2022}
}
@inproceedings{tang2020bipartite,
title={Bipartite Graph Reasoning GANs for Person Image Generation},
author={Tang, Hao and Bai, Song and Torr, Philip HS and Sebe, Nicu},
booktitle={BMVC},
year={2020}
}
If you use the original XingGAN, GestureGAN, C2GAN, and SelectionGAN model, please cite the following papers:
XingGAN
@inproceedings{tang2020xinggan,
title={XingGAN for Person Image Generation},
author={Tang, Hao and Bai, Song and Zhang, Li and Torr, Philip HS and Sebe, Nicu},
booktitle={ECCV},
year={2020}
}
GestureGAN
@article{tang2019unified,
title={Unified Generative Adversarial Networks for Controllable Image-to-Image Translation},
author={Tang, Hao and Liu, Hong and Sebe, Nicu},
journal={IEEE Transactions on Image Processing (TIP)},
year={2020}
}
@inproceedings{tang2018gesturegan,
title={GestureGAN for Hand Gesture-to-Gesture Translation in the Wild},
author={Tang, Hao and Wang, Wei and Xu, Dan and Yan, Yan and Sebe, Nicu},
booktitle={ACM MM},
year={2018}
}
C2GAN
@article{tang2021total,
title={Total Generate: Cycle in Cycle Generative Adversarial Networks for Generating Human Faces, Hands, Bodies, and Natural Scenes},
author={Tang, Hao and Sebe, Nicu},
journal={IEEE Transactions on Multimedia (TMM)},
year={2021}
}
@inproceedings{tang2019cycleincycle,
title={Cycle In Cycle Generative Adversarial Networks for Keypoint-Guided Image Generation},
author={Tang, Hao and Xu, Dan and Liu, Gaowen and Wang, Wei and Sebe, Nicu and Yan, Yan},
booktitle={ACM MM},
year={2019}
}
SelectionGAN
@article{tang2022multi,
title={Multi-channel attention selection gans for guided image-to-image translation},
author={Tang, Hao and Torr, Philip HS and Sebe, Nicu},
journal={Transactions on Pattern Analysis and Machine Intelligence (TPAMI)},
year={2022}
}
@inproceedings{tang2019multi,
title={Multi-channel attention selection gan with cascaded semantic guidance for cross-view image translation},
author={Tang, Hao and Xu, Dan and Sebe, Nicu and Wang, Yanzhi and Corso, Jason J and Yan, Yan},
booktitle={CVPR},
year={2019}
}
Contributions
If you have any questions/comments/bug reports, feel free to open a github issue or pull a request or e-mail to the author Hao Tang (bjdxtanghao@gmail.com).
Collaborations
I'm always interested in meeting new people and hearing about potential collaborations. If you'd like to work together or get in contact with me, please email bjdxtanghao@gmail.com. Some of our projects are listed here.
Success is the sum of small efforts, repeated day in and day out.