Home

Awesome

Towards Efficient and Scale-Robust Ultra-High-Definition Image Demoiréing

Project Page | Dataset | Paper

PWC PWC

Towards Efficient and Scale-Robust Ultra-High-Definition Image Demoireing (ECCV 2022)
Xin Yu, Peng Dai, Wenbo Li, Lan Ma, Jiajun Shen, Jia Li, Xiaojuan Qi

Example 1 Gif 1

:hourglass_flowing_sand: To Do

:rocket: :rocket: :rocket: News:

Introduction

When photographing the contents displayed on the digital screen, an inevitable frequency aliasing between the camera’s color filter array (CFA) and the screen’s LCD subpixel widely exists. The captured images are thus mixed with colorful stripes, named moire patterns, which severely degrade images’ perceptual qualities. Although a plethora of dedicated demoiré methods have been proposed in the research community recently, yet is still far from achieving promising results in the real-world scenes. The key limitation of these methods is that they all only conduct research on low-resolution or synthetic images. However, with the rapid development of mobile devices, modern widely-used mobile phones typically allow users to capture 4K resolution (i.e.,ultra-high-definition) images, and thus the effectiveness of these methods on this practical scenes is not promised. In this work, we explore moire pattern removal for ultra-high-definition images. First, we propose the first ultra-high-definition demoiréing dataset (UHDM), which contains 5,000 real-world 4K resolution image pair, and conduct a benchmark study on the current state of the art. Then, we analyze limitations of the state of the art and summarize the key issue of them, i.e., not scale-robust. To address their deficiencies, we deliver a plug-and-play semantic-aligned scale-aware module which helps us to build a frustratingly simple baseline model for tackling 4K moire images. Our framework is easy to implement and fast for inference, achieving state-of-the-art results on four demoiréing datasets while being much more lightweight. We hope our investigation could inspire more future research in this more practical setting in image demoiréing.

<p align="center"><img src="./figures/cost.png" width="55%" ></p>

Environments

First you have to make sure that you have installed all dependencies. To do so, you can create an anaconda environment called esdnet using

conda env create -f environment.yaml
conda activate esdnet

Our implementation has been tested on one NVIDIA 3090 GPU with cuda 11.2.

Quick Test

Once you have installed all dependencies, you can try a quick test without downloading training dataset:

1. Download our pre-trained models:

We provide pre-trained models on four datasets, which can be downloaded through the following links:

Model NameTraining DatasetDownload Link
ESDNetUHDMuhdm_checkpoint.pth
ESDNet-LUHDMuhdm_large_checkpoint.pth
ESDNetFHDMifhdmi_checkpoint.pth
ESDNet-LFHDMifhdmi_large_checkpoint.pth
ESDNetTIP2018tip_checkpoint.pth
ESDNet-LTIP2018tip_large_checkpoint.pth
ESDNetLCDMoireaim_checkpoint.pth
ESDNet-LLCDMoireaim_large_checkpoint.pth

Or you can simply run the following command for automatic downloading:

bash scripts/download_model.sh

Then the checkpoints will be included in the folder pretrain_model/.

2. Test with your own images:

Change the configuration file ./demo_config/demo.yaml to fit your own setting, and then simply run:

python demo_test.py --config ./demo_config/demo.yaml

the output images will be included in the path depending on the flags SAVE_PREFIX and EXP_NAME in your configuration file.

Dataset

Data We provide the 4K dataset UHDM for you to evaluate a pretrained model or train a new model. To this end, you can download them here, or you can simply run the following command for automatic data downloading:

bash scripts/download_data.sh

Then the dataset will be available in the folder uhdm_data/.

Train

To train a model from scratch, simply run:

python train.py --config CONFIG.yaml

where you replace CONFIG.yaml with the name of the configuration file you want to use. We have included configuration files for each dataset under the folder config/.

For example, if you want to train our lightweight model ESDNet on UHDM dataset, run:

python train.py --config ./config/uhdm_config.yaml

Test

To test a model, you can also simply run:

python test.py --config CONFIG.yaml

where you need to specify the value of TEST_EPOCH in the CONFIG.yaml to evaluate a model trained after specific epochs, or you can also specify the value of LOAD_PATH to directly load a pre-trained checkpoint.

Results

Quantitative Results:

<p align="center"> <img src="./figures/quantitative_results.png" width="100%"> </p>

Extended link:

If you want to remove moire patterns in your video, you can try our CVPR 2022 work: VDMoire vdmoire

Citation

Please consider :grimacing: staring this repository and citing the following papers if you feel this repository useful.

@inproceedings{yu2022towards,
  title={Towards efficient and scale-robust ultra-high-definition image demoir{\'e}ing},
  author={Yu, Xin and Dai, Peng and Li, Wenbo and Ma, Lan and Shen, Jiajun and Li, Jia and Qi, Xiaojuan},
  booktitle={European Conference on Computer Vision},
  pages={646--662},
  year={2022},
  organization={Springer}
}

@inproceedings{dai2022video,
  title={Video Demoireing with Relation-Based Temporal Consistency},
  author={Dai, Peng and Yu, Xin and Ma, Lan and Zhang, Baoheng and Li, Jia and Li, Wenbo and Shen, Jiajun and Qi, Xiaojuan},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year={2022}
}

Contact

If you have any questions, you can email me (yuxin27g@gmail.com).