Home

Awesome

<h2 align="center">LoveDA: A Remote Sensing Land-Cover Dataset for Domain Adaptive Semantic Segmentation</h2> <h5 align="right">by <a href="https://junjue-wang.github.io/homepage/">Junjue Wang</a>, <a href="http://zhuozheng.top/">Zhuo Zheng</a>, Ailong Ma, Xiaoyan Lu, and <a href="http://rsidea.whu.edu.cn/">Yanfei Zhong</a></h5>

[Paper], [Video], [Dataset], [BibTeX], [Leaderboard-SEG], [Leaderboard-UDA]

<div align="center"> <img src="https://github.com/Junjue-Wang/resources/blob/main/LoveDA/LoveDA.jpg?raw=true"> <img src="https://github.com/Junjue-Wang/resources/blob/main/LoveDA/statics_diff.png?raw=true"> </div>

News

Highlights

  1. 5987 high spatial resolution (0.3 m) remote sensing images from Nanjing, Changzhou, and Wuhan
  2. Focus on different geographical environments between Urban and Rural
  3. Advance both semantic segmentation and domain adaptation tasks
  4. Three considerable challenges:
    • Multi-scale objects
    • Complex background samples
    • Inconsistent class distributions

Citation

If you use LoveDA in your research, please cite our NeurIPS2021 paper.

    @inproceedings{NEURIPS DATASETS AND BENCHMARKS2021_4e732ced,
         author = {Wang, Junjue and Zheng, Zhuo and Ma, Ailong and Lu, Xiaoyan and Zhong, Yanfei},
         booktitle = {Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks},
         editor = {J. Vanschoren and S. Yeung},
         pages = {},
         publisher = {Curran Associates, Inc.},
         title = {LoveDA: A Remote Sensing Land-Cover Dataset for Domain Adaptive Semantic Segmentation},
         url = {https://datasets-benchmarks-proceedings.neurips.cc/paper_files/paper/2021/file/4e732ced3463d06de0ca9a15b6153677-Paper-round2.pdf},
         volume = {1},
         year = {2021}
    }
    @dataset{junjue_wang_2021_5706578,
        author={Junjue Wang and Zhuo Zheng and Ailong Ma and Xiaoyan Lu and Yanfei Zhong},
        title={Love{DA}: A Remote Sensing Land-Cover Dataset for Domain Adaptive Semantic Segmentation},
        month=oct,
        year=2021,
        publisher={Zenodo},
        doi={10.5281/zenodo.5706578},
        url={https://doi.org/10.5281/zenodo.5706578}
    }

Dataset and Contest

The LoveDA dataset is released at <b>Zenodo</b> and <b>Baidu Drive</b> Code: 27vc

You can develop your models on Train and Validation sets.

Category labels: background ā€“ 1, building ā€“ 2, road ā€“ 3, water ā€“ 4, barren ā€“ 5,forest ā€“ 6, agriculture ā€“ 7. And the no-data regions were assigned 0 which should be ignored. The provided data loader will help you construct your pipeline.

Submit your test results on <b>LoveDA Semantic Segmentation Challenge</b>, <b>LoveDA Unsupervised Domain Adaptation Challenge</b>. You will get your Test scores smoothly.

Feel free to design your own models, and we are looking forward to your exciting results!

License

The owners of the data and of the copyright on the data are RSIDEA, Wuhan University. Use of the Google Earth images must respect the "Google Earth" terms of use. All images and their associated annotations in LoveDA can be used for academic purposes only, <font color="red"><b> but any commercial use is prohibited.</b></font>

<a rel="license" href="https://creativecommons.org/licenses/by-nc-sa/4.0/deed.en"> <img alt="ēŸ„čƆ共äŗ«č®øåÆåč®®" style="border-width:0" src="https://i.creativecommons.org/l/by-nc-sa/4.0/88x31.png" /></a>