Awesome
Learning to Predict Layout-to-image Conditional Convolutions for Semantic Image Synthesis
Xihui Liu, Guojun Yin, Jing Shao, Xiaogang Wang and Hongsheng Li.<br> Published in NeurIPS 2019.
Paper | Poster | Slides
Installation
Clone this repo.
git clone https://github.com/xh-liu/CC-FPSE.git
cd CC-FPSE/
This code requires PyTorch 1.1+ and python 3+. Please install dependencies by
pip install -r requirements.txt
The results reported in the paper is trained on 16 TITANX GPUs.
Dataset Preparation
Follow the dataset preparation process in SPADE.
Generating Images Using Pretrained Model
-
Download the pretrained models from Google Drive Folder, and extract it to 'checkpoints/'.
-
Generate images using the pretrained model with test_coco.sh, test_ade.sh, and test_cityscapes.sh.
-
The outputs images are stored at
./results/[type]_pretrained/
by default. You can view them using the autogenerated HTML file in the directory.
Training New Models
New models can be trained with train.sh. This is an example of training the model on one machine.
Citation
If you use this code for your research, please cite our papers.
@inproceedings{liu2019learning,
title={Learning to Predict Layout-to-image Conditional Convolutions for Semantic Image Synthesis},
author={Liu, Xihui and Yin, Guojun and Shao, Jing and Wang, Xiaogang and Li, Hongsheng},
booktitle={Advances in Neural Information Processing Systems},
year={2019}
}
Acknowledgments
This code borrows heavily from SPADE.