Home

Awesome

OpenAQ API

Slack Chat

Deploy

Overview

This repository contains the source code for the OpenAQ API, a publicly-accessible API that provides endpoints to query the real-time and historical air quality measurements on the OpenAQ platform.

[!NOTE]
This repository is for setting up and deploying the OpenAQ API. If you just wish to access the public API to query data from the OpenAQ platform, visit https://api.openaq.org or https://docs.openaq.org to learn more.

Platform Overview

openaq-fetch and openaq-fetch-lcs take care of fetching new data and writing to S3. Lambda functions defined in ingest/, then load data into the database, defined in openaq-db.

Getting started

The API code is all in openaq_api

Deployment is managed with Amazon Web Services (AWS) Cloud Development Kit (CDK).

Dependencies

Local Development Environment

In production, the OpenAQ API runs on AWS Lambda with the help of the mangum library. This allows the application to run in a serverless environment and take advantage of async Python and FastAPI routing. Despite the serverless deployment, running the API locally as a standard FastAPI application is largely unchanged, making local development much easier.

Settings

Settings can be loaded using .env files, and multiple files can be kept and used. The easiest way to manage multiple environment files is to add an extension describing your environment. For example, if I wanted to keep a production, staging and local environment, I would save them as .env.production, .env.staging and .env.local each with their own settings.

uvicorn

The easiest way to run the API locally is to use uvicorn. Make sure that you have your settings (.env) file setup. Once that is done, you can run the following from the openaq_api/openaq_api directory. Variables from the .env files can be overrode by setting them inline.

# Run using the default .env file
uvicorn main:app --reload

You can also specify which .env file to load by passing the ENV variable. This should not include the .env. prefix

# Run our production environment
ENV=production uvicorn main:app --reload

And you can always override variables by setting them inline. This is handy for when you want to change something for the purpose of debugging.

# Run the staging environment and add verbose logging
ENV=staging LOG_LEVEL=debug uvicorn main:app --reload

Setting up your environment

Local development environment

To set up your environment, create a .env file that includes the following variables

AWS_PROFILE=openaq-prod
OPENAQ_APIUSER_PW=
OPENAQ_RWUSER_PW=
PGUSER=
PGPASSWORD=
PGDATABASE=
PGHOST=
PGPORT=


DATABASE_URL="postgres://apiuser:${OPENAQ_APIUSER_PW}@${PGHOST}:${PGPORT}/${PGDATABASE}?sslmode=require"
DATABASE_WRITE_URL="postgres://rwuser:${OPENAQ_RWUSER_PW}@${PGHOST}:${PGPORT}/${PGDATABASE}?sslmode=require"


# Prod
OPENAQ_ENV='staging'
OPENAQ_FASTAPI_URL=""

OPENAQ_FETCH_BUCKET=openaq-fetches
OPENAQ_ETL_BUCKET=openaq-fetches

AWS Cloud Deployment Environment

Additional environmnet variables are required for a full deployment to the AWS Cloud.

Rate limiting

In the production environment, rate limiting is handled in two places, AWS WAF and at the application level with Starlette Middleware. The application rate limiting is configurable via environment variables. The rate limiting middleware requires access to an instance of a redis cluster. For local development, docker can be a convenient method to set up a local redis cluster. With docker, use the following command:

docker run -e "IP=0.0.0.0" -p 7000-7005:7000-7005 grokzen/redis-cluster:7.0.7

Now a redis instance will be available at http://localhost:7000. Configure the REDIS_HOST to localhost and REDIS_PORT to 7000.

[!TIP] On some macOS systems port 7000 is used by Airplay which can complicate the mapping of ports from the Docker container. The easiest option is to disable the Airplay reciever in system settings. System settings -> General -> Airplay receiver (toggle off)

Rate limiting values

Rate limiting can be toggled off for local develop via the RATE_LIMITING environment variable. Other rate limiting values are:

e.g. RATE_AMOUNT_KEY=5 and RATE_TIME=1 would allow 5 requests per 1 minute.

[!NOTE] With AWS WAF, rate limiting also occurs at the cloudfront stage. The application level rate limiting should be less than or equal to the value set at AWS WAF.

Contributing

There are many ways to contribute to this project; more details can be found in the contributing guide.