Home

Awesome

Derain Methods

A Decoder-free Transformer-like Architecture for High-efficiency Single Image Deraining (IJCAI Long Oral, 3.7%)

Xiao Wu, Ting-Zhu Huang, Liang-Jian Deng, Tian-Jing Zhang

Paper | Video

Abstract: Despite the success of vision Transformers for the image deraining task, they are limited by computation-heavy and slow runtime. In this work, we investigate Transformer decoder is not necessary and has huge computational costs. Therefore, we revisit the standard vision Transformer as well as its successful variants and propose a novel Decoder-Free Transformer-Like (DFTL) architecture for fast and accurate single image deraining. Specifically, we adopt a cheap linear projection to represent visual information with lower computational costs than previous linear projections. Then we replace standard Transformer decoder block with designed Progressive Patch Merging (PPM), which attains comparable performance and efficiency. DFTL could significantly alleviate the computation and GPU memory requirements through proposed modules. Extensive experiments demonstrate the superiority of DFTL compared with competitive Transformer architectures, e.g., ViT, DETR, IPT, Uformer, and Restormer.

<hr />

Toy Example

image

Training and Evaluation

Training and Testing for Deraining:

<table> <tr> <th align="left">Derain</th> <th align="center">Dataset</th> <th align="center">Visual Results</th> </tr> <tr> <td align="left">Rain200L</td> <td align="center"><a href="https://www.icst.pku.edu.cn/struct/Projects/joint_rain_removal.html">Link</a></td> <td align="center"><a href="">Download</a></td> </tr> <tr> <td align="left">Rain200H</td> <td align="center"><a href="https://www.icst.pku.edu.cn/struct/Projects/joint_rain_removal.html">Link</a></td> <td align="center"><a href="">Download</a></td> </tr> <tr> <td>DID</td> <td align="center"><a href="https://github.com/hezhangsprinter/DID-MDN">Link</a></td> <td align="center"><a href="">Download</a></td> </tr> <tr> <td>DDN</td> <td align="center"><a href="https://xueyangfu.github.io/projects/cvpr2017.html">Link</a></td> <td align="center"><a href="">Download</a></td> </tr> </table>

Step1.

|-$ROOT/data
├── Rain200H
│   ├── train_c
│   │   ├── norain-1.png
│   │   ├── ...
│   ├── test_c
│   │   │   ├── norain-1.png
│   │   │   ├── ...

Step2. Open codes in your ide, run the following code:

python run_derain.py

Let's start to run our DFTL.

Currently, you can run run_DFTLW.py or run_DFTLX.py in Link.

Note: [Ready to use] Our project is based on MMCV, but you needn't to install it and master MMCV. More importantly, it can be more easy to introduce more methods.

Benchmark

We provide simple pipelines to train/test/inference models for a quick start.

<details open> <summary>Derain model zoo: </summary> </details>

Citation

If it is helpful for you, please kindly cite our paper:

  @inproceedings{DFTL,
    title     = {A Decoder-free Transformer-like Architecture for High-efficiency Single Image Deraining},
    author    = {Wu, Xiao and Huang, Ting-Zhu and Deng, Liang-Jian and Zhang, Tian-Jing},
    booktitle = {Proceedings of the Thirty-First International Joint Conference on
                 Artificial Intelligence (IJCAI-22)},
    pages     = {1474--1480},
    year      = {2022},
    month     = {7},
    doi       = {10.24963/ijcai.2022/205},
  }

Contact

Should you have any question, please contact wxwsx1997@gmail.com;

Acknowledgment: This code is based on the MMCV toolbox and Restormer.

Our Related Works

License & Copyright

This project is open sourced under GNU General Public License v3.0.