Home

Awesome

@fastify/under-pressure

CI NPM version neostandard javascript style

Measure process load with automatic handling of "Service Unavailable" plugin for Fastify. It can check maxEventLoopDelay, maxHeapUsedBytes, maxRssBytes and maxEventLoopUtilization values. You can also specify a custom health check, to verify the status of external resources.

<a name="requirements"></a>

Requirements

Fastify ^4.0.0. Please refer to this branch and related versions for Fastify ^1.1.0 compatibility.

<a name="install"></a>

Install

npm i @fastify/under-pressure

<a name="usage"></a>

Usage

Require the plugin and register it into the Fastify instance.

const fastify = require('fastify')()

fastify.register(require('@fastify/under-pressure'), {
  maxEventLoopDelay: 1000,
  maxHeapUsedBytes: 100000000,
  maxRssBytes: 100000000,
  maxEventLoopUtilization:0.98
})

fastify.get('/', (request, reply) => {
  if (fastify.isUnderPressure()) {
    // skip complex computation
  }
  reply.send({ hello: 'world'})
})

fastify.listen({ port: 3000 }, err => {
  if (err) throw err
  console.log(`server listening on ${fastify.server.address().port}`)
})

@fastify/under-pressure will automatically handle for you the Service Unavailable error once one of the thresholds has been reached. You can configure the error message and the Retry-After header.

fastify.register(require('@fastify/under-pressure'), {
  maxEventLoopDelay: 1000,
  message: 'Under pressure!',
  retryAfter: 50
})

You can also configure custom Error instance @fastify/under-pressure will throw.

class CustomError extends Error {
  constructor () {
    super('Custom error message')
    Error.captureStackTrace(this, CustomError)
  }
}

fastify.register(require('@fastify/under-pressure'), {
  maxEventLoopDelay: 1000,
  customError: CustomError
})

The default value for maxEventLoopDelay, maxHeapUsedBytes, maxRssBytes and maxEventLoopUtilization is 0. If the value is 0 the check will not be performed.

Thanks to the encapsulation model of Fastify, you can selectively use this plugin in some subset of routes or even with different thresholds in different plugins.

memoryUsage

This plugin also exposes a function that will tell you the current values of heapUsed, rssBytes, eventLoopDelay and eventLoopUtilized.

console.log(fastify.memoryUsage())

Pressure Handler

You can provide a pressure handler in the options to handle the pressure errors. The advantage is that you know why the error occurred. Moreover, the request can be handled as if nothing happened.

const fastify = require('fastify')()
const underPressure = require('@fastify/under-pressure')()

fastify.register(underPressure, {
  maxHeapUsedBytes: 100000000,
  maxRssBytes: 100000000,
  pressureHandler: (request, reply, type, value) => {
    if (type === underPressure.TYPE_HEAP_USED_BYTES) {
      fastify.log.warn(`too many heap bytes used: ${value}`)
    } else if (type === underPressure.TYPE_RSS_BYTES) {
      fastify.log.warn(`too many rss bytes used: ${value}`)
    }

    reply.send('out of memory') // if you omit this line, the request will be handled normally
  }
})

It is possible as well to return a Promise that will call reply.send (or something else).

fastify.register(underPressure, {
  maxHeapUsedBytes: 100000000,
  pressureHandler: (request, reply, type, value) => {
    return getPromise().then(() => reply.send({ hello: 'world' }))
  }
})

Any other return value than a promise or nullish will be sent to client with reply.send.

It's also possible to specify the pressureHandler on the route:

const fastify = require('fastify')()
const underPressure = require('@fastify/under-pressure')()

fastify.register(underPressure, {
  maxHeapUsedBytes: 100000000,
  maxRssBytes: 100000000,
})

fastify.register(async function (fastify) {
  fastify.get('/', {
    config: {
      pressureHandler: (request, reply, type, value) => {
        if (type === underPressure.TYPE_HEAP_USED_BYTES) {
          fastify.log.warn(`too many heap bytes used: ${value}`)
        } else if (type === underPressure.TYPE_RSS_BYTES) {
          fastify.log.warn(`too many rss bytes used: ${value}`)
        }

        reply.send('out of memory') // if you omit this line, the request will be handled normally
      }
    }
  }, () => 'A')
})

Status route

If needed you can pass { exposeStatusRoute: true } and @fastify/under-pressure will expose a /status route for you that sends back a { status: 'ok' } object. This can be useful if you need to attach the server to an ELB on AWS for example.

If you need the change the exposed route path, you can pass { exposeStatusRoute: '/alive' } options.

To configure the endpoint more specifically you can pass an object. This consists of

fastify.register(require('@fastify/under-pressure'), {
  maxEventLoopDelay: 1000,
  exposeStatusRoute: {
    routeOpts: {
      logLevel: 'debug',
      config: {
        someAttr: 'value'
      }
    },
    routeSchemaOpts: { // If you also want to set a custom route schema
      hide: true
    },
    url: '/alive' // If you also want to set a custom route path and pass options
  }
})

The above example will set the logLevel value for the /alive route to be debug.

If you need to return other information in the response, you can return an object from the healthCheck function (see next paragraph) and use the routeResponseSchemaOpts property to describe your custom response schema (note: status will always be present in the response)

fastify.register(underPressure, {
  ...
  exposeStatusRoute: {
    routeResponseSchemaOpts: {
      extraValue: { type: 'string' },
      metrics: {
        type: 'object',
        properties: {
          eventLoopDelay: { type: 'number' },
          rssBytes: { type: 'number' },
          heapUsed: { type: 'number' },
          eventLoopUtilized: { type: 'number' },
        },
      },
      // ...
    }
  },
  healthCheck: async (fastifyInstance) => {
    return {
      extraValue: await getExtraValue(),
      metrics: fastifyInstance.memoryUsage(),
      // ...
    }
  },
}

Custom health checks

If needed you can pass a custom healthCheck property, which is an async function, and @fastify/under-pressure will allow you to check the status of other components of your service.

This function should return a promise that resolves to a boolean value or to an object. The healthCheck function can be called either:

By default when this function is supplied your service health is considered unhealthy, until it has started to return true.

const fastify = require('fastify')()

fastify.register(require('@fastify/under-pressure'), {
  healthCheck: async function (fastifyInstance) {
    // do some magic to check if your db connection is healthy, etc...
    return true
  },
  healthCheckInterval: 500
})

<a name="sample-interval"></a>

Sample interval

You can set a custom value for sampling the metrics returned by memoryUsage using the sampleInterval option, which accepts a number that represents the interval in milliseconds.

The default value is different depending on which Node version is used. In version 8 and 10 it is 5, while on version 11.10.0 and up it is 1000. This difference is because from version 11.10.0 the event loop delay can be sampled with monitorEventLoopDelay and this allows to increase the interval value.

const fastify = require('fastify')()

fastify.register(require('@fastify/under-pressure'), {
  sampleInterval: <your custom sample interval in ms>
})

<a name="additional-information"></a>

Additional information

<a name="set-timeout-vs-set-interval"></a>

setTimeout vs setInterval

Under the hood the @fastify/under-pressure uses the setTimeout method to perform its polling checks. The choice is based on the fact that we do not want to add additional pressure to the system.

In fact, it is known that setInterval will call repeatedly at the scheduled time regardless of whether the previous call ended or not, and if the server is already under load, this will likely increase the problem, because those setInterval calls will start piling up. setTimeout, on the other hand, is called only once and does not cause the mentioned problem.

One note to consider is that because the two methods are not identical, the timer function is not guaranteed to run at exactly the same rate when the system is under pressure or running a long-running process.

<a name="acknowledgements"></a>

Acknowledgements

This project is kindly sponsored by LetzDoIt.

<a name="license"></a>

License

Licensed under MIT.