Home

Awesome

Satellite Image Land Cover Segmentation using U-net

This GitHub repository is developed by Srimannarayana Baratam and Georgios Apostolides as a part of Computer Vision by Deep Learning (CS4245) course offered at TU Delft. The implementation of the code was done using PyTorch, it uses U-net architecture to perform multi-class semantic segmentation. The repository from which our implementation has been derived can be found [here]. A well articulated blog is also available [here] for the project by the authors of this repository.

Google Colab Wrapper

For testing the repository, a google colab wrapper is also provided which explains in detail how to execute the code along with insights. Just download the "colab_wrapper.ipynb" file from the repository and open in your colab. Instructions are available there to clone this repository directly to your drive and train using GPU runtime.

Dataset

The dataset we used is taken from the DeepGlobe Challenge of Land Cover Segmentation in 2018. [DeepGlobeChallenges] However, the server for the challenge is no longer available for submission and evaluation of solutions. [DeepGlobe2018 Server] and the validation and test set are not accompanied by labels. For this reason we are using only the training set of the challenge and we are further splitting it into validation and test set to be able to evaluate our solution. The original dataset can be downloaded from Kaggle [DeepGLobe2018 Original Dataset] here and the dataset we use can be downloaded from [Link] separated into the training/validation and test set we used for our model.

Files Explanation

In this section we will present the different files inside the repository as well as an explanation about their functionality

File NameExplanation / Function
U-net_markdown.ipynb<img width=90/>Used as a wrapper to run the train and predict scripts.
train.pyUsed for the training of the network.
predict.pyUsed to perform a prediction on the test dataset.
diceloss.pyIt's a function used to calculate the dice loss.
classcount.pyCalculates the weights to be used for the weighted cross entropy by counting the pixel number of each class in the train dataset.
distribution.pyUsed to evaluate the pixel numbers of the validation and training set and visualize them via bar chart.
dataset.pyUsed as a data loader during the training phase.

Training

The following flags can be used while training the model.

<ins>Guidelines<ins>

-f : Used to load a model already stored in memory.
-e : Used to specify the Number of training epochs.
-l : Used to specify the learning rate to be used for training.
-b : Used to specify the batch size.
-v : Used to specify the percentage of the validation split (1-100).
-s : Used to specify the scale of the image to be used for training.

<ins>Example:<ins/>

Training the model for 100 epochs using 20% of the data as a validation split, learning rate is 4x10^-5, batch size is 2 and image scale is 20%

!python3 train.py -e 100 -v 20.0 -l 4e-5 -b 2 -s 0.2

Prediction

<ins>Guidelines<ins>

-m : Used to specify the directory to the model.
-i : Used to specify the directory of the images to test the model on.
-o : Used to specify the directory in which the predictions will be outputted.
-s : Used to specify the scale of the images to be used for predictions.
--viz: Saves the predictions in the form of an image.
(For best results used the same scale you used for training the model)

Note: Inference of scale 0.2 takes approximately 10 minutes.

<ins>Example<ins>

Making a prediction on the full test set dataset using 30 epoch model trained on full data using a scale of 20%. The script outputs the IoU score of the model.

%%time
!python predict.py -m data/checkpoints/model_ep30_full_data.pth -i data/<test_set_directory>/* -o predictions/ -s 0.2 --viz