Home

Awesome

<img align="right" height="39" src="https://github.com/EL-BID/Building-Detection/blob/master/img/azavea_RGB_72dpi_trans_sm.png"><img align="right" width="115" height="49" src="https://github.com/EL-BID/Building-Detection/blob/master/img/IDB_logo.jpg">

Raster Vision Building Detection

Description and Context

This repo contains code for running a Raster Vision experiment to train a model to detect buildings from satellite imagery in three cities in Latin America:

The experiment analyzes the imagery, generates training chips, trains a model, make predictions, evaluates the predictions, and produces a prediction package which can be used to make predictions on new imagery. Training the model in a reasonable amount of time (a day as opposed to a month) requires a machine with a high-end NVIDIA GPU. It's possible to rent one in the cloud using the instructions for using AWS Batch. Below is an example of predictions over Georgetown, with predicted building pixels colored orange.

Example predictions over Georgetown

Installation Guide

Docker setup

First, install Docker (preferably version 18 or above). These instructions assume a Unix-based environment; however, it should also be possible to run the Docker container in Mac or Windows environments.

This experiment was tested with Raster Vision commit 9f38cc9. The corresponding Docker images are:

To use this with a more recent commit/version of Raster Vision it may be necessary to update the experiment configuration accordingly.

Clone repo and download data

To get the scripts, source files, and data needed to run the experiment, clone this repo and then retrieve the data bundle from this link.

Unzip the data bundle on the local filesystem to run the experiment locally. To run the experiment on AWS, unzip the dataset and upload it to an S3 bucket. The expected layout of this data bundle is

├── input
│   ├── belice
│   │   ├── aois
│   │   ├── imagery
│   │   ├── test-aois
│   │   └── train-aois
│   ├── georgetown
│   │   ├── aois
│   │   ├── imagery
│   │   ├── test-aois
│   │   └── train-aois
│   ├── paramaribo
│   │   ├── aois
│   │   ├── imagery
│   │   ├── test-aois
│   │   └── train-aois
│   └── vector-tiles
│       ├── belize.mbtiles.gz
│       ├── guyana.mbtiles.gz
│       └── suriname.mbtiles.gz
├── predictions
│   └── command-config.json
└── results
    ├── belice
    │   ├── belice-1-polygons.geojson
    │   ├── belice.tif
    │   └── screenshots
    ├── georgetown
    │   ├── georgetown-1-polygons.geojson
    │   ├── georgetown.tif
    │   └── screenshots
    ├── paramaribo
    │   ├── paramaribo-1-polygons.geojson
    │   ├── paramaribo.tif
    │   └── screenshots
    └── predict-package
        ├── georgetown_sample.tif
        ├── out_sample.tif
        └── predict_package.zip

The input directory contains everything necessary for running the experiment. For each city there is imagery and a set of training and test AOIs in GeoJSON format. There are also vector tiles for each country which are used for extracting labels. The results directory has predictions (GeoTIFF and GeoJSON) and screenshots for each city, and the prediction package. Set the IDB_DATA_DIR environment variable to point to the local data bundle using

> export IDB_DATA_DIR=<data bundle dir>

Before running experiments on AWS, or changing the layout of the data bundle, update the constants pointing to the data and RV output in idb/experiment.py#L7-14 Note that the local variables (local_rv_root and local_data_root) should be prefixed with /opt/data since that is the directory in the Docker container that the IDB_DATA_DIR is mounted to. There is no need to update the local variables if the layout of the data bundle has not changed.

Users's Guide

Run the Docker console

Raster Vision commands need to be run from inside the Raster Vision Docker container, which can be accessed using:

> scripts/console

This will mount the following directories from the local filesystem into the container:

Run experiment locally

The experiment in this repo trains a Mobilenet DeepLab model (pretrained on Pascal VOC) on small training AOIs over three large ~50k x ~50k scenes for 150k steps with a batch size of 8. The labels come from Open Street Map vector tiles. The predictions cover the entire scenes; therefore, the predictions over the training AOIs should be ignored. The evaluation metrics are computed using labels from held-out validation AOIs.

The commands to run an experiment must be invoked inside the Docker console, ie. after running ./scripts/console. Before running a full experiment, it is a good idea to run a small test experiment to check that things are setup correctly using:

> rastervision run local -e idb.experiment -a test True

The main thing to check here is that it didn't crash. The model will not be not be useful since it was only trained for one step.

Run experiment on a local GPU

To run Raster Vision on a local GPU, run the console with the --gpu flag. This will use the GPU version of the container and the nvidia Docker runtime. A full experiment can then be run with:

> rastervision run local -e idb.experiment

Run experiment on an AWS GPU

To run an experiment on AWS, setup a Raster Vision Batch stack on your AWS account by following the instructions in the Raster Vision AWS repository. Using the names of the GPU and CPU Batch job definition and job queue that were created (which can be seen in the Batch console), set the appropriate configuration in your Raster Vision profile in $HOME/.rastervision/default which should contain:

[AWS_BATCH]
job_queue=<gpu job queue>
job_definition=<gpu job definition>
cpu_job_queue=<cpu job queue>
cpu_job_definition=<cpu job definition>

When running ./scripts/console, set the AWS_PROFILE environment variable to the desired profile, and use the --aws flag which will forward your AWS credentials and profile.

To run a full experiment remotely on AWS use:

> rastervision run aws_batch -e idb.experiment -a remote True

On a P3.2xlarge instance, this should take about a day.

Viewing Tensorboard

During training, results can be monitored using Tensorboard at either localhost:6006/ if running locally, or <public_dns>:6006 if running on AWS, where <public_dns> is the public DNS of the EC2 instance that is running the training job. To forward these ports to your localhost, run the console with the --tensorboard flag.

Inspect output

The output of the experiment will be in local_rv_root or remote_rv_root (the variables set here) depending on whether it was run locally or remotely. The layout of the output directory is documented here. The eval.json file generated by the eval command that is part of the experiment should contain metrics similar to those in sample-eval.json, with a building F1 score around 0.68.

Make predictions on new imagery

The bundle command that was run as part of the experiment generates a prediction package which can be used to make predictions on new imagery. To make predictions using a sample image and the prediction package that comes with the data bundle use:

> rastervision predict /opt/data/results/predict-package/predict_package.zip /opt/data/results/predict-package/georgetown_sample.tif /opt/data/output.tif

The predictions, visualized in QGIS can be seen at the top of this file. Note that the predict command does not generate polygon output, only rasters. The polygon output can only be produced at the moment by the predict stage when running an experiment.

Customizing the experiment

The experiment in this repo is configured to work with a specific set of scenes, but these can be modified by editing scenes-config.json and the splits variable in idb/experiment.py#L74-87. To find the training and validation AOIs, we viewed imagery along with an OSM basemap in QGIS and then created polygons for well-labeled areas. If there are no well-labeled areas, you may need to contribute annotations to OSM, or choose a different city. The vector tile .mbtiles files included with the data bundle and used in this experiment were downloaded from OSM QA Tiles Country Extracts. Since these files can be very large, it is advisable to create a smaller extract focused on specific city using a tool like Osmium.

For more general customization of the experiment, see the Raster Vision experiment config docs.

Note on problematic GeoTIFFs

Raster Vision considers zero valued pixels to be NODATA and skips over them when making predictions. Unfortunately, the GeoTIFF for Belice that was provided by IDB has a problem. The parts of the scene that were not captured by the satellite that appear black when viewed in QGIS (that ought to have NODATA values) contain very low values (0-6) for unknown reasons. Because of this, Raster Vision makes many false positive predictions over these regions.

To get around this problem, there is a script at idb/nodata_transform.py which generates a new GeoTIFF where values below a minimum are set to zero. The experiment in this repo uses the fixed version of the GeoTIFF. When running an experiment on new imagery, if there are false positives over what looks like a NODATA part of the scene, this may be the problem, and using the nodata_transform script may help. For example, the Belice file was corrected using

python -m idb.nodata_transform 10 /opt/data/input/belice/imagery/belice.tif \
    /opt/data/input/belice/imagery/new_belice.tif

Authors

Lewis Fishgold @ Azavea

License

The Documentation of Support and Use of the software is licensed under Creative Commons IGO 3.0 Attribution-NonCommercial-NoDerivative (CC-IGO 3.0 BY-NC-ND)

The codebase of this repo uses AM-331-A3 Software License.

Credits

This application uses Open Source components. You can find the source code of these components along with license information below. We acknowledge and are grateful to these developers for their contributions to open source.

Project: Raster Vision Source Code: https://github.com/azavea/raster-vision License: Apache 2.0

Project: Rasterio Source Code: https://github.com/mapbox/rasterio License: Mapbox

Project: Numpy Source Code: https://github.com/numpy/numpy License: Numpy

Project: Click Source Code: https://github.com/pallets/click License: Click