Home

Awesome

Augmentation of rPPG Benchmark Datasets: Learning to Remove and Embed rPPG Signals via Double Cycle Consistent Learning from Unpaired Facial Videos

This repository is the official implementation of Augmentation of rPPG Benchmark Datasets: Learning to Remove and Embed rPPG Signals via Double Cycle Consistent Learning from Unpaired Facial Videos that has been accepted to ECCV 2022.

Description

We propose a RErPPG-Net to augment existing rPPG datasets by embedding ground-truth PPG signals into any existing facial videos.

The proposed RErPPG-Net consists of a Removal-Net $G_{R}$ and an Embedding-Net $G_{E}$ and aims to remove any inherent rPPG signals existing in the input videos and then to embed the specified PPG signals into the rPPG-removed videos.

<img src="pipeline.PNG" width="800">

Implementation

The RErPPG-Net and the rPPG estimator were trained with Nvidia RTX 2080 and RTX 3080.

Training

The RErPPG-Net / rPPG estimator were trained with 900 / 500 epochs.

Optimizer : Adam optimizer with the learning rate of 0.001.

Batch size : RErPPG-Net / rPPG estimator were trained with 1 / 3 batch size.

Dataset

To generate the Aug-rPPG dataset, we use all the 76 training videos and the corresponding PPG signals from UBFC-rPPG and PURE datasets as the inputs to the proposed RErPPG-Net. The 76 input videos are from 42 subjects, where 35 subjects are from UBFC-rPPG training set and 7 subjects are from PURE training set.

By running every possible combination of the videos and PPG signals, we generate 5776 videos of resolution 200*200 pixels. Note that, because we only include the facial region of 200X200 pixels in the data augmentation, our generated videos are of the same quality as the two benchmark datasets.

Becauew of privacy issues, we can not directly release the "Aug-rPPG" dataset. Please contact nthumplab740@gmail.com, we will reply the download link of "Aug-rPPG".

Contributing

If you find this work useful, consider citing our work using the following bibTex: