Home

Awesome

CST

Code release for "Cycle Self-Training for Domain Adaptation" (NeurIPS 2021)

Prerequisites

Training

VisDA-2017

CUDA_VISIBLE_DEVICES=0 python run_cst.py data/visda-2017 -d VisDA2017 -s Synthetic -t Real -a resnet101 \
--epochs 30 --early 12 --lr 0.002 --per-class-eval --temperature 3.0 --center-crop --log logs/cst/VisDA2017 \
--trade-off 0.08 trade-off1 2.0 --trade-off3 0.5 --threshold 0.97 -b 28 

Office Home

CUDA_VISIBLE_DEVICES=0 python run_cst.py data/office-home -d OfficeHome -s Pr -t Rw -a resnet50 \
--epochs 30 --early 30 --temperature 2.5 --bottleneck-dim 2048 --log logs/cst/OfficeHome_Pr2Rw \
--trade-off1 2.0 --trade-off3 0.5 --threshold 0.97 --trade-off 0.015

Acknowledgement

This code is implemented based on the Transfer Learning Library, and it is our pleasure to acknowledge their contributions.

The SAM code is adapted from https://github.com/davda54/sam.

Citation

If you use this code for your research, please consider citing:

@article{liu2021cycle,
  title={Cycle Self-Training for Domain Adaptation},
  author={Liu, Hong and Wang, Jianmin and Long, Mingsheng},
  journal={arXiv preprint arXiv:2103.03571},
  year={2021}
}

Contact

If you have any problem about our code, feel free to contact