Awesome
SEA-T2F
Multi-caption Text-to-Face Synthesis: Database and Algorithm
Requirements: python=3.7, pytorch=1.1.0, torchvision=0.3.0
- Download data:
Download the CelebAText-HQ from https://drive.google.com/drive/folders/1IAb_iy6-soEGQWhbgu6cQODsIUJZpahC?usp=sharing, and extract them into data/CelebAText-HQ.
Download the Multi-modal CelebA-HQ from https://github.com/IIGROUP/Multi-Modal-CelebA-HQ-Dataset and extract them into data/Multi-modal.
Download the CelebAMask-HQ from https://github.com/switchablenorms/CelebAMask-HQ and copy CelebAimg into data/CelebAText-HQ or data/Multi-modal CelebA-HQ.
- Download pretrain model:
Download the pretriain model of CelebAText-HQ from https://drive.google.com/drive/folders/1XufcOo_I09h86ZR2M4UJ8WTVR7ARF87d, and extract them into /DAMSMEencoders/CelebAText-HQ.
Download the pretriain model of Multi-modal CelebA-HQ from https://drive.google.com/drive/folders/1FN4q7xD1jKvXeG3Pd6wqieUlr52bVoIX, and extract them into /DAMSMEencoders/CelebAText-HQ.
- Download generative model:
Download the generative model of CelebAText-HQ from https://drive.google.com/drive/folders/1XufcOo_I09h86ZR2M4UJ8WTVR7ARF87d, and copy it into /models.
Download the generative model of Multi-modal CelebA-HQ from https://drive.google.com/drive/folders/1FN4q7xD1jKvXeG3Pd6wqieUlr52bVoIX, and copy it into /models.
- Train:
python main.py --cfg cfg/train_face.yml --gpu 7
- Evaluation:
python main.py --cfg cfg/eval_face.yml --gpu 7