Home

Awesome

HDTF

Flow-guided One-shot Talking Face Generation with a High-resolution Audio-visual Dataset <a href="https://openaccess.thecvf.com/content/CVPR2021/papers/Zhang_Flow-Guided_One-Shot_Talking_Face_Generation_With_a_High-Resolution_Audio-Visual_Dataset_CVPR_2021_paper.pdf" target="_blank">paper</a> <a href="https://github.com/MRzzm/HDTF/blob/main/Supplementary%20Materials.pdf" target="_blank">supplementary</a>

Details of HDTF dataset

./HDTF_dataset consists of youtube video url, video resolution (in our method, may not be the best resolution), time stamps of talking face, facial region (in the our method) and the zoom scale of the cropped window. xx_video_url.txt:

format:     video name | video youtube url

xx_resolution.txt:

format:    video name | resolution(in our method)

xx_annotion_time.txt:

format:    video name | time stamps of clip1 | time stamps of clip2 | time stamps of clip3....

xx_crop_wh.txt:

format:    video name+clip index | min_width | width |  min_height | height (in our method)

xx_crop_ratio.txt:

format:    video name+clip index | window zoom scale

Processing of HDTF dataset

When using HDTF dataset,

The HDTF dataset is available to download under a <a href="https://creativecommons.org/licenses/by/4.0/" target="_blank"> Creative Commons Attribution 4.0 International License</a>. If you face any problems when processing HDTF, pls contact me.

Downloading

For convenience, we added the download.py script which downloads, crops and resizes the dataset. You can use it via the following command:

python download.py --output_dir /path/to/output/dir --num_workers 8

Note: some videos might become unavailable if the authors will remove them or make them private.

Reference

if you use HDTF, pls reference

@inproceedings{zhang2021flow,
  title={Flow-Guided One-Shot Talking Face Generation With a High-Resolution Audio-Visual Dataset},
  author={Zhang, Zhimeng and Li, Lincheng and Ding, Yu and Fan, Changjie},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={3661--3670},
  year={2021}
}