Home

Awesome

<h1 align="center"> <img src="https://media.licdn.com/dms/image/D4E0BAQFh9M8zrNFjQw/company-logo_200_200/0/1665352677142?e=2147483647&v=beta&t=iI4YPPusisIuK3I-VjYnt7WVuIA4jsSQpglIwfr9X2U" alt="windmill" width="100"> </h1> <h4 align="center">Turn scripts into workflows and UIs in minutes</h4> <p align="center"> <img src="https://github.com/windmill-labs/windmill-helm-charts/actions/workflows/helm_test.yml/badge.svg" alt="drawing"/> <img src="https://github.com/windmill-labs/windmill-helm-charts/actions/workflows/release.yml/badge.svg" alt="drawing"/> <img src="https://img.shields.io/github/v/release/windmill-labs/windmill-helm-charts" alt="drawing"/> <img src="https://img.shields.io/github/downloads/windmill-labs/windmill-helm-charts/total.svg" alt="drawing"/> </p> <p align="center"> <a href="#install">Install</a> • <a href="#core-values">Core Values</a> • <a href="#full-values">Full Values</a> • <a href="#local-s3">Local S3</a> • <a href="#caveats">Caveats</a> • <a href="#kubernetes-hosting-tips">Kubernetes Tips</a> </p> <hr>

Install

Have Helm 3 installed.

helm repo add windmill https://windmill-labs.github.io/windmill-helm-charts/
helm install mywindmill windmill/windmill -n windmill --create-namespace --values values.yaml

To update versions:

helm repo update windmill
helm upgrade mywindmill windmill/windmill -n windmill --values values.yaml

You do not need to provide a values.yaml to be able to test it on minikube. Follow the steps below.

When using a non super-user role for postgresql in databaseUrl

You will need to setup some required roles which would otherwise be done automatically when using a super-user role for the databaseUrl.

Follow those instructions

Test it on minikube

To make it work on a local minkube to test. Get the ip address of the ingress:

▶ kubectl get ingress -n windmill
NAME       CLASS    HOSTS                        ADDRESS        PORTS   AGE
windmill   <none>   windmill,windmill,windmill   192.168.49.2   80      13m

If not ip address is displayed, enable the ingress addon:

minikube addons enable ingress

Then modify /etc/hosts to match the baseDomain, by default 'windmill'.

E.g:

192.168.49.2   windmill

Then open your browser at http://windmill

Even if you setup oauth, login as** admin@windmill.dev **/ changeme to setup the instance & accounts and give yourself admin privileges.

Core Values

# windmill root values block
windmill:
  # domain as shown in browser, this is used together with `baseProtocol` as part of the BASE_URL environment variable in app and worker container and in the ingress resource, if enabled
  baseDomain: windmill
  baseProtocol: http
  # postgres URI, pods will crashloop if database is unreachable, sets DATABASE_URL environment variable in app and worker container
  databaseUrl: postgres://postgres:windmill@windmill-postgresql/windmill?sslmode=disable
  # replica for the application app
  appReplicas: 2
  # replicas for the workers, jobs are executed on the workers
  lspReplicas: 2
  workerGroups:
    # The default worker group is the one that will execute jobs with any taggs  except the native ones. Windmill has a default worker group configuration for it
    - name: "default"
      replicas: 3
      # -- Annotations to apply to the pods
      annotations: {}

      # -- Labels to apply to the pods
      labels: {}

      # -- Node selector to use for scheduling the pods
      nodeSelector: {}

      # -- Tolerations to apply to the pods
      tolerations: []

      # -- Affinity rules to apply to the pods
      affinity: {}

      # -- Resource limits and requests for the pods
      resources:
        requests:
          memory: "1028Mi"
          cpu: "500m"
        limits:
          memory: "2048Mi"
          cpu: "1000m"

      # -- Extra environment variables to apply to the pods
      extraEnv: []

      # -- Extra sidecar containers
      extraContainers: []

      # -- Mode for workers, defaults to "worker" - alternative "agent" requires Enterprise license
      mode: "worker"

    # Thenative worker group will only execute native jobs. Windmill has a default worker group configuration for it
    - name: "native"
      replicas: 4
      # -- Resource limits and requests for the pods
      resources:
        requests:
          memory: "128Mi"
          cpu: "100m"
        limits:
          memory: "256Mi"
          cpu: "200m"

      # -- Extra environment variables to apply to the pods
      extraEnv: []

      # -- Extra sidecar containers
      extraContainers: []

      # -- Mode for workers, defaults to "worker" - alternative "agent" requires Enterprise license
      mode: "worker"

    - name: "gpu"
      replicas: 0

  # Use those to override the tag or image used for the app and worker containers. Windmill uses the same image for both.
  # By default, if enterprise is enable, the image is set to ghcr.io/windmill-labs/windmill-ee, otherwise the image is set to ghcr.io/windmill-labs/windmill
  #tag: "mytag"
  #image: "ghcr.io/windmill-labs/windmill"

# enable postgres (bitnami) on kubernetes
postgresql:
  enabled: true

# enable minio (bitnami) on kubernetes
minio:
  enabled: false

# Configure Ingress
# ingress:
#   className: ""

# enable enterprise features
enterprise:
  # -- enable windmill enterprise, requires license key.
  enabled: false

Full Values

KeyTypeDefaultDescription
enterprise.enabledboolfalseenable Windmill Enterprise , requires license key.
enterprise.enabledS3DistributedCacheboolfalse
enterprise.licenseKeystring"123456F"Windmill provided Enterprise license key. Sets LICENSE_KEY environment variable in app and worker container.
enterprise.nsjailboolfalseuse nsjail for sandboxing
enterprise.s3CacheBucketstring"mybucketname"S3 bucket to use for dependency cache. Sets S3_CACHE_BUCKET environment variable in worker container
enterprise.samlMetadatastring""SAML Metadata URL to enable SAML SSO (Can be set in the Instance Settings UI, which is the recommended method)
enterprise.scimTokenstring""
ingress.annotationsobject{}
ingress.classNamestring""
ingress.enabledbooltrueenable/disable included ingress resource
ingress.tlslist[]TLS config for the ingress resource. Useful when using cert-manager and nginx-ingress
minio.auth.rootPasswordstring"windmill"
minio.auth.rootUserstring"windmill"
minio.enabledboolfalseenabled included Minio operator for s3 resource demo purposes
minio.fullnameOverridestring"windmill-minio"
minio.modestring"standalone"
minio.primary.enabledbooltrue
postgresql.auth.databasestring"windmill"
postgresql.auth.postgresPasswordstring"windmill"
postgresql.enabledbooltrueenabled included Postgres container for demo purposes only using bitnami
postgresql.fullnameOverridestring"windmill-postgresql"
postgresql.primary.persistence.enabledbooltrue
serviceAccount.annotationsobject{}
serviceAccount.createbooltrue
serviceAccount.namestring""
windmill.app.affinityobject{}Affinity rules to apply to the pods
windmill.app.annotationsobject{}Annotations to apply to the pods
windmill.app.autoscaling.enabledboolfalseenable or disable autoscaling
windmill.app.autoscaling.maxReplicasint10maximum autoscaler replicas
windmill.app.autoscaling.targetCPUUtilizationPercentageint80target CPU utilization
windmill.app.extraEnvlist[]Extra environment variables to apply to the pods
windmill.app.labelsobject{}Labels to apply to the pods
windmill.app.nodeSelectorobject{}Node selector to use for scheduling the pods
windmill.app.resourcesobject{}Resource limits and requests for the pods
windmill.app.tolerationslist[]Tolerations to apply to the pods
windmill.appReplicasint2replica for the application app
windmill.baseDomainstring"windmill"domain as shown in browser, this variable and baseProtocol are used as part of the BASE_URL environment variable in app and worker container and in the ingress resource, if enabled
windmill.baseProtocolstring"http"protocol as shown in browser, change to https etc based on your endpoint/ingress configuration, this variable and baseDomain are used as part of the BASE_URL environment variable in app and worker container
windmill.cookieDomainstring""domain to use for the cookies. Use it if windmill is hosted on a subdomain and you need to share the cookies with the hub for instance
windmill.databaseUrlstring"postgres://postgres:windmill@windmill-postgresql/windmill?sslmode=disable"Postgres URI, pods will crashloop if database is unreachable, sets DATABASE_URL environment variable in app and worker container
windmill.databaseUrlSecretNamestring""name of the secret storing the database URI, take precedence over databaseUrl. The key of the url is 'url'
windmill.denoExtraImportMapstring""custom deno extra import maps (syntax: key1=value1,key2=value2)
windmill.exposeHostDockerboolfalsemount the docker socket inside the container to be able to run docker command as docker client to the host docker daemon
windmill.globalErrorHandlerPathstring""if set, the path to a script in the admins workspace that will be triggered upon any jobs failure
windmill.imagestring""windmill image tag, will use the Acorresponding ee or ce image from ghcr if not defined. Do not include tag in the image name.
windmill.instanceEventsWebhookstring""send instance events to a webhook. Can be hooked back to windmill
windmill.lsp.affinityobject{}Affinity rules to apply to the pods
windmill.lsp.annotationsobject{}Annotations to apply to the pods
windmill.lsp.autoscaling.enabledboolfalseenable or disable autoscaling
windmill.lsp.autoscaling.maxReplicasint10maximum autoscaler replicas
windmill.lsp.autoscaling.targetCPUUtilizationPercentageint80target CPU utilization
windmill.lsp.extraEnvlist[]Extra environment variables to apply to the pods
windmill.lsp.labelsobject{}Labels to apply to the pods
windmill.lsp.nodeSelectorobject{}Node selector to use for scheduling the pods
windmill.lsp.resourcesobject{}Resource limits and requests for the pods
windmill.lsp.tagstring"latest"
windmill.lsp.tolerationslist[]Tolerations to apply to the pods
windmill.lspReplicasint2replicas for the workers, jobs are executed on the workers
windmill.multiplayer.affinityobject{}Affinity rules to apply to the pods
windmill.multiplayer.annotationsobject{}Annotations to apply to the pods
windmill.multiplayer.autoscaling.enabledboolfalseenable or disable autoscaling
windmill.multiplayer.autoscaling.maxReplicasint10maximum autoscaler replicas
windmill.multiplayer.autoscaling.targetCPUUtilizationPercentageint80target CPU utilization
windmill.multiplayer.extraEnvlist[]Extra environment variables to apply to the pods
windmill.multiplayer.labelsobject{}Labels to apply to the pods
windmill.multiplayer.nodeSelectorobject{}Node selector to use for scheduling the pods
windmill.multiplayer.resourcesobject{}Resource limits and requests for the pods
windmill.multiplayer.tagstring"latest"
windmill.multiplayer.tolerationslist[]Tolerations to apply to the pods
windmill.multiplayerReplicasint1replicas for the lsp containers used by the app
windmill.npmConfigRegistrystring""pass the npm for private registries
windmill.pipExtraIndexUrlstring""pass the extra index url to pip for private registries
windmill.pipIndexUrlstring""pass the index url to pip for private registries
windmill.pipTrustedHoststring""pass the trusted host to pip for private registries
windmill.rustLogstring"info"rust log level, set to debug for more information etc, sets RUST_LOG environment variable in app and worker container
windmill.tagstring""windmill app image tag, will use the App version if not defined
windmill.workerGroups[0].affinityobject{}Affinity rules to apply to the pods
windmill.workerGroups[0].annotationsobject{}Annotations to apply to the pods
windmill.workerGroups[0].extraEnvlist[]Extra environment variables to apply to the pods
windmill.workerGroups[0].extraContainerslist[]Extra containers as sidecars
windmill.workerGroups[0].labelsobject{}Labels to apply to the pods
windmill.workerGroups[0].namestring"default"
windmill.workerGroups[0].nodeSelectorobject{}Node selector to use for scheduling the pods
windmill.workerGroups[0].replicasint3
windmill.workerGroups[0].resourcesobject{"limits":{"cpu":"1000m","memory":"2048Mi"},"requests":{"cpu":"500m","memory":"1028Mi"}}Resource limits and requests for the pods
windmill.workerGroups[0].tolerationslist[]Tolerations to apply to the pods
windmill.workerGroups[0].modestring"worker"Mode for workers, "worker" or "agent", agent requires Enterprise
windmill.workerGroups[0].commandlist[]Command to run, overrides image default command
windmill.workerGroups[1].affinityobject{}Affinity rules to apply to the pods
windmill.workerGroups[1].annotationsobject{}Annotations to apply to the pods
windmill.workerGroups[1].extraEnvlist[]Extra environment variables to apply to the pods
windmill.workerGroups[1].extraContainerslist[]Extra containers as sidecars
windmill.workerGroups[1].labelsobject{}Labels to apply to the pods
windmill.workerGroups[1].namestring"gpu"
windmill.workerGroups[1].nodeSelectorobject{}Node selector to use for scheduling the pods
windmill.workerGroups[1].replicasint0
windmill.workerGroups[1].resourcesobject{}Resource limits and requests for the pods
windmill.workerGroups[1].tolerationslist[]Tolerations to apply to the pods
windmill.workerGroups[1].modestring"worker"Mode for workers, "worker" or "agent", agent requires Enterprise
windmill.workerGroups[1].commandlist[]Command to run, overrides image default command
windmill.workerGroups[2].affinityobject{}Affinity rules to apply to the pods
windmill.workerGroups[2].annotationsobject{}Annotations to apply to the pods
windmill.workerGroups[2].extraEnvlist[]Extra environment variables to apply to the pods
windmill.workerGroups[2].extraContainerslist[]Extra containers as sidecars
windmill.workerGroups[2].labelsobject{}Labels to apply to the pods
windmill.workerGroups[2].namestring"native"
windmill.workerGroups[2].nodeSelectorobject{}Node selector to use for scheduling the pods
windmill.workerGroups[2].replicasint4
windmill.workerGroups[2].resourcesobject{"limits":{"cpu":"200m","memory":"256Mi"},"requests":{"cpu":"100m","memory":"128Mi"}}Resource limits and requests for the pods
windmill.workerGroups[2].tolerationslist[]Tolerations to apply to the pods
windmill.workerGroups[2].modestring"worker"Mode for workers, "worker" or "agent", agent requires Enterprise
windmill.workerGroups[2].commandlist[]Command to run, overrides image default command

Local S3

The chart includes a Minio S3 distribution to demonstrate the usage of S3 as a resource in a vendor-agnostic environment like Kubernetes. The local Minio S3 service will be available to the Windmill workers through its Kubernetes service, which is set to "windmill-minio" by default. In the Resources page, you should create an S3 API Connection Object, and import it as a connection object to reduce code duplication between scripts. For the sake of this example, this stage is skipped. Below is an example of how to authenticate and use the provided local S3 distribution in a Python script running in Windmill:

from minio import Minio

def main():
    # Create a client with the MinIO server, its access key
    # and secret key.
    client = Minio(
        "windmill-minio", # Local Kubernetes Service
        access_key="windmill",
        secret_key="windmill",
    )

    # Make 'demo' bucket if not exist.
    found = client.bucket_exists("demo")
    if not found:
        client.make_bucket("demo")
    else:
        print("Bucket 'demo' already exists")

    with open('readme.txt', 'w') as f:
        f.write('Create a new text file!')

    client.fput_object(
        "demo", "readme.txt", "readme.txt",
    )

    print(
        "'readme.txt' is successfully uploaded as "
        "object 'readme.txt' to bucket 'demo'."
    )

Enterprise features

To use the enterprise version with the <license key> provided upon subscription, add the following to the values.yaml file:

enterprise:
  enabled: true

Then go to the superadmin settings -> instance settings -> license key and set your license key

S3 Cache

Enterprise users can use S3 storage for dependency caching for performance reasons at high scale (use only with #workers > 20). Cache is two way synced at regular intervals (10 minutes). To use it, the worker deployment requires access to an S3 bucket. There are several ways to do this:

  1. On AWS (and EKS) , you can use a service account with IAM roles attached. See AWS docs - once you have a policy , you can create an account via eksctl for instance

    eksctl create iamserviceaccount --name serviceaccountname --namespace production --cluster windmill-cluster --role-name "iamrolename" \ --attach-policy-arn arn:aws:iam::12312315:policy/bucketpolicy --approve
    
  2. Mount/attach a credentials file in /root/.aws/credentials of the worker deployment

  3. Add environment variables for the AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY, via kube secrets.

The sync relies on rclone and uses its methods of authentication to s3 per Rclone documentation

Then the values settings become:

enterprise:
  enabled: true
  enabledS3DistributedCache: true
  s3CacheBucket: mybucketname

Caveats

Kubernetes Hosting Tips

Ingress configuration

The helm chart does have an ingress configuration included. It's enabled by default. The ingress uses the windmill.baseDomain variable for its hostname configuration. Here are example configurations for a few cloud providers.

It configures the HTTP ingress for the app, lsp and multiplayer containers. The configuration (except for plain nginx ingress) also exposes the windmill app SMTP service for email triggers on a separate IP address/domain name. This is the IP address/domain name you need to point your MX/A records to, learn more here.

AWS ALB

windmill:
  baseDomain: "windmill.example.com"
  app:
    smtpService:
      enabled: true
      annotations:
        service.beta.kubernetes.io/aws-load-balancer-type: "external"
        service.beta.kubernetes.io/aws-load-balancer-nlb-target-type: "ip"
        service.beta.kubernetes.io/aws-load-balancer-scheme: "internet-facing"
        # # for static ip (more info on https://kubernetes-sigs.github.io/aws-load-balancer-controller/v2.7/guide/service/annotations/#eip-allocations):
        # service.beta.kubernetes.io/aws-load-balancer-eip-allocations: eipalloc-xxxxxxxxxxxxxxxxx,eipalloc-yyyyyyyyyyyyyyyyy
    ...
  ...
...

ingress:
  className: "alb"
  annotations:
    alb.ingress.kubernetes.io/scheme: internet-facing
    alb.ingress.kubernetes.io/tags: Environment=dev,Team=test
    alb.ingress.kubernetes.io/target-type: ip
    alb.ingress.kubernetes.io/target-group-attributes: stickiness.enabled=true,stickiness.lb_cookie.duration_seconds=604800,stickiness.type=lb_cookie
    alb.ingress.kubernetes.io/load-balancer-attributes: idle_timeout.timeout_seconds=600
    alb.ingress.kubernetes.io/group.name: windmill
    alb.ingress.kubernetes.io/group.order: '10'
    alb.ingress.kubernetes.io/certificate-arn: my-certificatearn

GCP + GCE LB + managed certificates

windmill:
  baseDomain: "windmill.example.com"
  app:
    service:
      annotations:
        cloud.google.com/backend-config: '{"default": "session-config"}'
    smtpService:
      enabled: true
      annotations:
        cloud.google.com/l4-rbs: "enabled"
        # # for static ip (more info on https://cloud.google.com/kubernetes-engine/docs/concepts/service-load-balancer-parameters#spd-static-ip-parameters):
        # networking.gke.io/load-balancer-ip-addresses: <REGIONAL_IP_NAME>
  lsp:
    service:
      annotations:
        cloud.google.com/backend-config: '{"default": "session-config"}'
        
ingress:
  annotations:
    kubernetes.io/ingress.class: "gce"
    kubernetes.io/ingress.global-static-ip-name: <GLOBAL_IP_NAME>
    networking.gke.io/managed-certificates: managed-cert

Replace <GLOBAL_IP_NAME> with the name of a global static IP address you've created in GCP.

In addition to the above, you will need to apply the following resources for session affinity and managed certificates:

apiVersion: cloud.google.com/v1
kind: BackendConfig
metadata:
  name: windmill-backendconfig
spec:
  sessionAffinity:
    affinityType: "GENERATED_COOKIE"
    affinityCookieTtlSec: 86400 # max
apiVersion: networking.gke.io/v1
kind: ManagedCertificate
metadata:
  name: managed-cert
spec:
  domains:
    - windmill.example.com

Azure + app routing + keyvault certificate

windmill:
  baseDomain: "windmill.example.com"
  app:
    smtpService:
      enabled: true
      # # for static ip (more info on https://learn.microsoft.com/en-us/azure/aks/static-ip):
      # annotations:
      #   service.beta.kubernetes.io/azure-pip-name: <myAKSPublicIP>
ingress:
  annotations:
    kubernetes.azure.com/tls-cert-keyvault-uri: <KeyVaultCertificateUri>
  className: webapprouting.kubernetes.azure.com
  tls:
    - hosts:
      - "windmill.example.com"
      secretName: keyvault-windmill

You can find more details about SSL certificates with webapprouting in Azure here.

NGINX ingress + cert-manager:

windmill:
  baseDomain: "windmill.example.com"
  ...

ingress:
  className: "nginx"
  tls:
    - hosts:
        - "windmill.example.com"
      secretName: windmill-tls-cert
  annotations:
    cert-manager.io/issuer: "letsencrypt-prod"
    nginx.ingress.kubernetes.io/affinity: "cookie"
    nginx.ingress.kubernetes.io/affinity-mode: "persistent"
    nginx.ingress.kubernetes.io/session-cookie-name: "route"
...

You will also need to install cert-manager and configure an issuer. More details here and here. Cert-manager can also be used with the other cloud providers.

Generic

There are many ways to expose an app and it will depend on the requirements of your environment. If you don't want to use the included ingress and roll your own, you can just disable it. Overall, you want the following endpoints accessible included in the chart:

If you are using Prometheus and if the enterprise edition is enabled, you can scrape the windmill-app-metrics service on port 8001 at /metrics endpoint to gather stats about the Windmill application.

Docker dind configruation

If you don't want to run Docker using the host's Docker engine, you can use Docker-in-Docker (dind). Below is the configuration:

windmill:
  workerGroups:
  - name: "native"
    replicas: 2
    volumes:
    - emptyDir: {}
      name: sock-dir
    - emptyDir: {}
      name: windmill-workspace
    volumeMounts:
    - mountPath: /var/run
      name: sock-dir
    - mountPath: /opt/windmill
      name: windmill-workspace
    extraContainers:
    - args:
      - --mtu=1450
      image: docker:27.2.1-dind
      imagePullPolicy: IfNotPresent
      name: dind
      resources: {}
      securityContext:
        privileged: true
      terminationMessagePath: /dev/termination-log
      terminationMessagePolicy: File
      volumeMounts:
      - mountPath: /opt/windmill
        name: windmill-workspace
      - mountPath: /var/run
        name: sock-dir
    mode: "worker"

NOTE: the windmill-workspace volumeMount is used to share files between the dind container and the worker container.

<!--