Home

Awesome

<div align="center"> <h1>Underwater Image Enhancement by Diffusion Model with Customized CLIP-Classifier</h1> <div> <h4 align="center"> <a href="https://oucvisiongroup.github.io/CLIP-UIE.html/" target='_blank'>[Project Page]</a>ā€¢ <a href="" target='_blank'>[arXiv]</a> </h4> </div> <table> <tr> <td><center><img src="images/overflow.jpg" height="300">

The preparation for the pre-trained model. (a) Randomly select template A from the template pool (underwater domain). Then, the Color Transfer module, guided by template A, degrades an in-air natural image from INaturalist into underwater domain, constructing paired datasets for training image-to-image diffusion model. (b) The image-to-image diffusion model SR3 is trained to learn the prior knowledge, the mapping from the real underwater degradation domain and the real in-air natural domain, and to generate the corresponding enhancement results under the condition of the input synthetic underwater images produced by Color Transfer. </center></td>

</tr> </table> </div>

:desktop_computer: Requirements

:running_woman: Inference

šŸ“¦ Models

NameModel
CLIP-UIEDownload šŸ”—
Learned PromptDownload šŸ”—

Testing steps:

Thanks

Our code is based on SR3 and CLIP-LIT. You can refer to their README files and source code for more implementation details.

:love_you_gesture: Citation

If you find our work useful for your research, please consider citing the paper: