Awesome
Accelerating Image Super-Resolution Networks with Pixel-Level Classification (ECCV 2024)
<div align="justify"> <b>Abstract</b>: In recent times, the need for effective super-resolution (SR) techniques has surged, especially for large-scale images ranging 2K to 8K resolutions. For DNN-based SISR, decomposing images into overlapping patches is typically necessary due to computational constraints. In such patch-decomposing scheme, one can allocate computational resources differently based on each patch's difficulty to further improve efficiency while maintaining SR performance. However, this approach has a limitation: computational resources is uniformly allocated within a patch, leading to lower efficiency when the patch contain pixels with varying levels of restoration difficulty. To address the issue, we propose the Pixel-level Classifier for Single Image Super-Resolution (PCSR), a novel method designed to distribute computational resources adaptively at the pixel level. A PCSR model comprises a backbone, a pixel-level classifier, and a set of pixel-level upsamplers with varying capacities. The pixel-level classifier assigns each pixel to an appropriate upsampler based on its restoration difficulty, thereby optimizing computational resource usage. Our method allows for performance and computational cost balance during inference without re-training. Our experiments demonstrate PCSR's advantage over existing patch-distributing methods in PSNR-FLOP trade-offs across different backbone models and benchmarks. </div> <br>Dependencies
- Python 3.7<br>
- Pytorch 1.13<br>
- NVIDIA GPU + CUDA<br>
- Python packages:
pip install numpy opencv-python pandas tqdm fast_pytorch_kmeans
Demo
We offer a demo that can simulate 4x super-resolution using a single input image:
python demo.py --img_path <image path> --k <value> [--adaptive] [--no_refinement]
For a detailed description of the arguments and types of output, please refer to demo.py
and the results
directory.
Train
We offer a variety of pretrained models available in the save
directory.<br>
Data preprocessing
We use the DIV2K training set as our base.
cd datasets/scripts
python data_augmentation.py
python extract_subimages.py
python generate_mod_LR_bic.py
Run training scripts (example usage)
(Please move the pretrained model to a different directory to train a new model!)
# original model
./dist.sh train.py --config configs/carn-x4.yaml --gpu 0
# PCSR model (2-stage)
./dist.sh train_pcsr.py --config configs/carn-pcsr-phase0.yaml --gpu 0,1
./dist.sh train_pcsr.py --config configs/carn-pcsr-phase1.yaml --gpu 0
Test
Please refer to test.py
and test_pcsr.py
for a detailed description of the arguments.
# original model
python test.py --config <config path> --hr_data <hr foler> --lr_data <lr folder> --per_image --crop
# PCSR model (phase1)
python test_pcsr.py --config <config path> --hr_data <hr foler> --scale <integer scale> --per_image --crop --k 0 [--adaptive]
Citation
@misc{jeong2024acceleratingimagesuperresolutionnetworks,
title={Accelerating Image Super-Resolution Networks with Pixel-Level Classification},
author={Jinho Jeong and Jinwoo Kim and Younghyun Jo and Seon Joo Kim},
year={2024},
eprint={2407.21448},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2407.21448},
}