Home

Awesome

DSSL

Here's the official repo for Directional Self-supervised Learning for Heavy Image Augmentations [CVPR2022]

Folder contents:

Usage:

img.png

Overview of standard self-supervised learning and our DSSL. Original standard image transformations generate the standard views, and the harder view is derived from the the standard view by applying heavy augmentation RandAugment. (a) Standard instance- wise learning with standard views. (b) Instance-wise self-supervised learning after introducing heavily augmented (harder) views. Applying symmetric loss to maximize the similarity between standard and heavily augmented views roughly expands the feature cluster in the visual embedding space. The model may confuse the instance-level identity. (c) DSSL: To prevent the adverse effect from missing information of heavily augmented views, DSSL avoids arbitrarily maximizing their visual agreement. To tighten the feature cluster, DSSL applies an asymmetric loss for only gathering each heavily augmented view to its relevant standard view.

img_3.png

Illustration of our directional self-supervised learn- ing (DSSL) framework. Left: Construction of partially ordered views. Right: Symmetric loss LS for bidirectionally maximizing the agreement among augmented view pairs sampled from VT remains same. Asymmetric loss LA is proposed for encouraging the representations of the heavily augmented views to be close to their source standard views, respecting the partially ordered rela- tionship in VT → ̂ T

img_1.png

Comparisons on linear evaluation accuracies (%). repro: our reproduction of each method. collapse: model collapsed during training. w/ ̂ T : training views are jointly augmented from standard and heavy augmentations. Heavy augmentations perform unstably even model collapsing, while DSSL consistently benefits from ̂ T

img_2.png

Linear evaluation accuracies (%) of BYOL by applying more heavy augmentations (+ ̂T ).