Home

Awesome

QS-Attn: Query-Selected Attention for Contrastive Learning in I2I Translation (CVPR2022)

arXiv arXiv video

Unpaired image-to-image (I2I) translation often requires to maximize the mutual information between the source and the translated images across different domains, which is critical for the generator to keep the source content and prevent it from unnecessary modifications. The self-supervised contrastive learning has already been successfully applied in the I2I. By constraining features from the same location to be closer than those from different ones, it implicitly ensures the result to take content from the source. However, previous work uses the features from random locations to impose the constraint, which may not be appropriate since some locations contain less information of source domain. Moreover, the feature itself does not reflect the relation with others. This paper deals with these problems by intentionally selecting significant anchor points for contrastive learning. We design a query-selected attention (QS-Attn) module, which compares feature distances in the source domain, giving an attention matrix with a probability distribution in each row. Then we select queries according to their measurement of significance, computed from the distribution. The selected ones are regarded as anchors for contrastive loss. At the same time, the reduced attention matrix is employed to route features in both domains, so that source relations maintain in the synthesis. We validate our proposed method in three different I2I datasets, showing that it increases the image quality without adding learnable parameters.

<p align="center"> <img src="./teaser.png" width="800px"/> <br></p> <p align="center"> <img src="./arch.png" width="800px"/> <br> QS-Attn applies attention to select anchors for contrastive learning in single-direction I2I task </p>

Getting Started

Prerequisites

Pretrained Models

We provide Global, Local and Global+Local models for three datasets.

ModelCityscapesHorse2zebraAFHQ
GlobalCityscapes_GlobalHorse2zebra_GlobalAFHQ_Global
LocalCityscapes_LocalHorse2zebra_LocalAFHQ_Local
Global+LocalCityscapes_Global+LocalHorse2zebra_Global+LocalAFHQ_Global+Local

Training

bash ./datasets/download_qsattn_dataset.sh horse2zebra
python train.py \
--dataroot=datasets/horse2zebra \
--name=horse2zebra_global \
--QS_mode=global

Inference

python test.py \
--dataroot=datasets/horse2zebra \
--name=horse2zebra_qsattn_global \
--QS_mode=global

Citation

If you use this code for your research, please cite

@inproceedings{hu2022qs,
  title={QS-Attn: Query-Selected Attention for Contrastive Learning in I2I Translation},
  author={Hu, Xueqi and Zhou, Xinyue and Huang, Qiusheng and Shi, Zhengyi and Sun, Li and Li, Qingli},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={18291--18300},
  year={2022}
}