Home

Awesome

OSTeC: One-Shot Texture Complition

[CVPR 2021]

arXiv Prepring License: CC BY-NC-SA 4.0

Baris Gecer <sup> 1,2</sup>, Jiankang Deng <sup> 1,2</sup>, & Stefanos Zafeiriou <sup> 1,2</sup> <br/> <sup>1 </sup>Imperial College London <br/> <sup>2 </sup> Huawei CBG <br/>

<p align="center"> <img width="12%" src="figures/gifs/im18.gif" style="background-color:white; display: inline;" /> <img width="12%" src="figures/gifs/norah.gif" style="background-color:white; display: inline;" /> <img width="12%" src="figures/gifs/robin.gif" style="background-color:white; display: inline;" /> <img width="12%" src="figures/gifs/teaser2.gif" style="background-color:white; display: inline;" /> <img width="12%" src="figures/gifs/teaser3.gif" style="background-color:white; display: inline;" /> <img width="12%" src="figures/gifs/teaser4.gif" style="background-color:white; display: inline;" /> <img width="12%" src="figures/gifs/teaset5.gif" style="background-color:white; display: inline;" /> </p> <p align="center"><img width="100%" src="figures/teaser.jpg" style="background-color:white;" /></p>

Abstract

The last few years have witnessed the great success of non-linear generative models in synthesizing high-quality photorealistic face images. Many recent 3D facial texture reconstruction and pose manipulation from a single image approaches still rely on large and clean face datasets to train image-to-image Generative Adversarial Networks (GANs). Yet the collection of such a large scale high-resolution 3D texture dataset is still very costly and difficult to maintain age/ethnicity balance. Moreover, regression-based approaches suffer from generalization to the in-the-wild conditions and are unable to fine-tune to a target-image. In this work, we propose an unsupervised approach for one-shot 3D facial texture completion that does not require large-scale texture datasets, but rather harnesses the knowledge stored in 2D face generators. The proposed approach rotates an input image in 3D and fill-in the unseen regions by reconstructing the rotated image in a 2D face generator, based on the visible parts. Finally, we stitch the most visible textures at different angles in the UV image-plane. Further, we frontalize the target image by projecting the completed texture into the generator. The qualitative and quantitative experiments demonstrate that the completed UV textures and frontalized images are of high quality, resembles the original identity, can be used to train a texture GAN model for 3DMM fitting and improve pose-invariant face recognition.

Overview

<p align="center"><img width="100%" src="figures/overview.jpg" style="background-color:white;" /></p> Overview of the method. The proposed approach iteratively optimizes the texture UV-maps for different re-rendered images with their masks. At the end of each optimization, generated images are used to acquire partial UV images by dense landmarks. Finally, the completed UV images are fed to the next iteration for progressive texture building. <br/>

Requirements

This implementation is only tested under Ubuntu environment with Nvidia GPUs and CUDA 10.0 and CuDNN-7.0 installed.

Installation

1. Clone the repository and set up a conda environment as follows:

git clone https://github.com/barisgecer/OSTeC --recursive
cd OSTeC
conda env create -f environment.yml -n ostec
source activate ostec

2. Installation of Deep3DFaceRecon_pytorch

cd external/deep3dfacerecon/nvdiffrast    # ./OSTeC/external/deep3dfacerecon/nvdiffrast 
pip install .
cd ..    # ./OSTeC/external/deep3dfacerecon/
git clone https://github.com/deepinsight/insightface.git
cp -r ./insightface/recognition/arcface_torch/ ./models/
OSTeC
│
└─── external
     │
     └─── deep3dfacerecon
          │
          └─── BFM
              │
              └─── 01_MorphableModel.mat
              │
              └─── Exp_Pca.bin
              |
              └─── ...
OSTeC
│
└─── external
     │
     └─── deep3dfacerecon
          │
          └─── checkpoints
               │
               └─── face_recon
                   │
                   └─── epoch_latest.pth

3. Download Face Recognition & Landmark Detection & VGG & Style-Encoder models

OSTeC
│
└─── models
     │
     └─── resnet_18_20191231.h5
     │
     └─── vgg16_zhang_perceptual.pkl
     │
     └─── alignment
     │         .
     │         .
     │
     └─── fr_models
               .
               .

4. Download Face Segmentation models

OSTeC
│
└─── models
     │
     └─── Graphonomy
         │
         └─── inference.pth
<!--- ### 4. Download StyleGANv2 model - Download the model from the original repo: https://nvlabs-fi-cdn.nvidia.com/stylegan2/networks/stylegan2-ffhq-config-f.pkl And place it under 'models' directory like the following: ``` OSTeC │ └─── models │ └─── stylegan2_networks_stylegan2-ffhq-config-f ``` -->

Usage

More Results

<p align="center"><img width="100%" src="figures/comp2.jpg" /></p> <p align="center"><img width="100%" src="figures/comp1.jpg" /></p> <br/>

License

Acknowledgement

Citation

If you find this work is useful for your research, please cite our paper:

@InProceedings{Gecer_2021_CVPR,
    author    = {Gecer, Baris and Deng, Jiankang and Zafeiriou, Stefanos},
    title     = {OSTeC: One-Shot Texture Completion},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2021},
    pages     = {7628-7638}
}
<br/>