Awesome
<!-- markdownlint-disable --><a href="https://cpco.io/homepage"><img src=".github/banner.png?raw=true" alt="Project Banner"/></a><br/> <p align="right"> <a href="https://github.com/cloudposse-terraform-components/aws-eks-github-actions-runner/releases/latest"><img src="https://img.shields.io/github/release/cloudposse-terraform-components/aws-eks-github-actions-runner.svg?style=for-the-badge" alt="Latest Release"/></a><a href="https://slack.cloudposse.com"><img src="https://slack.cloudposse.com/for-the-badge.svg" alt="Slack Community"/></a></p>
<!-- markdownlint-restore --> <!-- ** DO NOT EDIT THIS FILE ** ** This file was automatically generated by the `cloudposse/build-harness`. ** 1) Make all changes to `README.yaml` ** 2) Run `make init` (you only need to do this once) ** 3) Run`make readme` to rebuild this file. ** ** (We maintain HUNDREDS of open source projects. This is how we maintain our sanity.) ** -->This component deploys self-hosted GitHub Actions Runners and a Controller on an EKS cluster, using "runner scale sets".
This solution is supported by GitHub and supersedes the actions-runner-controller developed by Summerwind and deployed by Cloud Posse's actions-runner-controller component.
Current limitations
The runner image used by Runner Sets contains
no more packages than are necessary
to run the runner. This is in contrast to the Summerwind implementation, which contains some commonly needed packages
like build-essential
, curl
, wget
, git
, and jq
, and the GitHub hosted images which contain a robust set of
tools. (This is a limitation of the official Runner Sets implementation, not this component per se.) You will need to
install any tools you need in your workflows, either as part of your workflow (recommended), by maintaining a
custom runner image,
or by running such steps in a
separate container that has the tools
pre-installed. Many tools have publicly available actions to install them, such as actions/setup-node
to install
NodeJS or dcarbone/install-jq-action
to install jq
. You can also install packages using
awalsh128/cache-apt-pkgs-action
, which has the advantage of being able to skip the installation if the package is
already installed, so you can more efficiently run the same workflow on GitHub hosted as well as self-hosted runners.
:::info
There are (as of this writing) open feature requests to add some commonly needed packages to the official Runner Sets runner image. You can upvote these requests here and here to help get them implemented.
:::
In the current version of this component, only "dind" (Docker in Docker) mode has been tested. Support for "kubernetes" mode is provided, but has not been validated.
Many elements in the Controller chart are not directly configurable by named inputs. To configure them, you can use the
controller.chart_values
input or create a resources/values-controller.yaml
file in the component to supply values.
Almost all the features of the Runner Scale Set chart are configurable by named inputs. The exceptions are:
- There is no specific input for specifying an outbound HTTP proxy.
- There is no specific input for supplying a custom certificate authority (CA) certificate to use when connecting to GitHub Enterprise Server.
You can specify these values by creating a resources/values-runner.yaml
file in the component and setting values as
shown by the default Helm
values.yaml,
and they will be applied to all runners.
Currently, this component has some additional limitations. In particular:
- The controller and all runners and listeners share the Image Pull Secrets. You cannot use different ones for different runners.
- All the runners use the same GitHub secret (app or PAT). Using a GitHub app is preferred anyway, and the single GitHub app serves the entire organization.
- Only one controller is supported per cluster, though it can have multiple replicas.
These limitations could be addressed if there is demand. Contact Cloud Posse Professional Services if you would be interested in sponsoring the development of any of these features.
Ephemeral work storage
The runners are configured to use ephemeral storage for workspaces, but the details and defaults can be a bit confusing.
When running in "dind" ("Docker in Docker") mode, the default is to use emptyDir
, which means space on the kubelet
base directory, which is usually the root disk. You can manage the amount of storage allowed to be used with
ephemeral_storage
requests and limits, or you can just let it use whatever free space there is on the root disk.
When running in kubernetes
mode, the only supported local disk storage is an ephemeral PersistentVolumeClaim
, which
causes a separate disk to be allocated for the runner pod. This disk is ephemeral, and will be deleted when the runner
pod is deleted. When combined with the recommended ephemeral runner configuration, this means that a new disk will be
created for each job, and deleted when the job is complete. That is a lot of overhead and will slow things down
somewhat.
The size of the attached PersistentVolume is controlled by ephemeral_pvc_storage
(a Kubernetes size string like "1G")
and the kind of storage is controlled by ephemeral_pvc_storage_class
(which can be omitted to use the cluster default
storage class).
This mode is also optionally available when using dind
. To enable it, set ephemeral_pvc_storage
to the desired size.
Leave ephemeral_pvc_storage
at the default value of null
to use emptyDir
storage (recommended).
Beware that using a PVC may significantly increase the startup of the runner. If you are using a PVC, you may want to keep idle runners available so that jobs can be started without waiting for a new runner to start.
Usage
Stack Level: Regional
Once the catalog file is created, the file can be imported as follows.
import:
- catalog/eks/github-actions-runner
...
The default catalog values e.g. stacks/catalog/eks/github-actions-runner.yaml
components:
terraform:
eks/github-actions-runner:
vars:
enabled: true
ssm_region: "us-east-2"
name: "gha-runner-controller"
charts:
controller:
chart_version: "0.7.0"
runner_sets:
chart_version: "0.7.0"
controller:
kubernetes_namespace: "gha-runner-controller"
create_namespace: true
create_github_kubernetes_secret: true
ssm_github_secret_path: "/github-action-runners/github-auth-secret"
github_app_id: "123456"
github_app_installation_id: "12345678"
runners:
config-default: &runner-default
enabled: false
github_url: https://github.com/cloudposse
# group: "default"
# kubernetes_namespace: "gha-runner-private"
create_namespace: true
# If min_replicas > 0 and you also have do-not-evict: "true" set
# then the idle/waiting runner will keep Karpenter from deprovisioning the node
# until a job runs and the runner is deleted.
# override by setting `pod_annotations: {}`
pod_annotations:
karpenter.sh/do-not-evict: "true"
min_replicas: 0
max_replicas: 8
resources:
limits:
cpu: 1100m
memory: 1024Mi
ephemeral-storage: 5Gi
requests:
cpu: 500m
memory: 256Mi
ephemeral-storage: 1Gi
self-hosted-default:
<<: *runner-default
enabled: true
kubernetes_namespace: "gha-runner-private"
# If min_replicas > 0 and you also have do-not-evict: "true" set
# then the idle/waiting runner will keep Karpenter from deprovisioning the node
# until a job runs and the runner is deleted. So we override the default.
pod_annotations: {}
min_replicas: 1
max_replicas: 12
resources:
limits:
cpu: 1100m
memory: 1024Mi
ephemeral-storage: 5Gi
requests:
cpu: 500m
memory: 256Mi
ephemeral-storage: 1Gi
self-hosted-large:
<<: *runner-default
enabled: true
resources:
limits:
cpu: 6000m
memory: 7680Mi
ephemeral-storage: 90G
requests:
cpu: 4000m
memory: 7680Mi
ephemeral-storage: 40G
Authentication and Secrets
The GitHub Action Runners need to authenticate to GitHub in order to do such things as register runners and pickup jobs. You can authenticate using either a GitHub App or a Personal Access Token (classic). The preferred way to authenticate is by creating and installing a GitHub App. This is the recommended approach as it allows for much more restricted access than using a Personal Access Token (classic), and the Action Runners do not currently support using a fine-grained Personal Access Token.
Site note about SSM and Regions
This component supports using AWS SSM to store and retrieve secrets. SSM parameters are regional, so if you want to deploy to multiple regions you have 2 choices:
- Create the secrets in each region. This is the most robust approach, but requires you to create the secrets in each region and keep them in sync.
- Create the secrets in one region and use the
ssm_region
input to specify the region where they are stored. This is the easiest approach, but does add some obstacles to managing deployments during a region outage. If the region where the secrets are stored goes down, there will be no impact on runners in other regions, but you will not be able to deploy new runners or modify existing runners until the SSM region is restored or until you set up SSM parameters in a new region.
Alternatively, you can create Kubernetes secrets outside of this component (perhaps using SOPS) and reference them by name. We describe here how to save the secrets to SSM, but you can save the secrets wherever and however you want to, as long as you deploy them as Kubernetes secret the runners can reference. If you store them in SSM, this component will take care of the rest, but the standard Terraform caveat applies: any secrets referenced by Terraform will be stored unencrypted in the Terraform state file.
Creating and Using a GitHub App
Follow the instructions here to create and install a GitHub App for the runners to use for authentication.
At the App creation stage, you will be asked to generate a private key. This is the private key that will be used to
authenticate the Action Runner. Download the file and store the contents in SSM using the following command, adjusting
the profile, region, and file name. The profile should be the terraform
role in the account to which you are deploying
the runner controller. The region should be the region where you are deploying the primary runner controller. If you are
deploying runners to multiple regions, they can all reference the same SSM parameter by using the ssm_region
input to
specify the region where they are stored. The file name (argument to cat
) should be the name of the private key file
you downloaded.
# Adjust profile name and region to suit your environment, use file name you chose for key
AWS_PROFILE=acme-core-gbl-auto-terraform AWS_REGION=us-west-2 chamber write github-action-runners github-auth-secret -- "$(cat APP_NAME.DATE.private-key.pem)"
You can verify the file was correctly written to SSM by matching the private key fingerprint reported by GitHub with:
AWS_PROFILE=acme-core-gbl-auto-terraform AWS_REGION=us-west-2 chamber read -q github-action-runners github-auth-secret | openssl rsa -in - -pubout -outform DER | openssl sha256 -binary | openssl base64
At this stage, record the Application ID and the private key fingerprint in your secrets manager (e.g. 1Password). You may want to record the private key as well, or you may consider it sufficient to have it in SSM. You will need the Application ID to configure the runner controller, and want the fingerprint to verify the private key. (You can see the fingerprint in the GitHub App settings, under "Private keys".)
Proceed to install the GitHub App in the organization or repository you want to use the runner controller for, and record the Installation ID (the final numeric part of the URL, as explained in the instructions linked above) in your secrets manager. You will need the Installation ID to configure the runner controller.
In your stack configuration, set the following variables, making sure to quote the values so they are treated as strings, not numbers.
github_app_id: "12345"
github_app_installation_id: "12345"
OR (obsolete): Creating and Using a Personal Access Token (classic)
Though not recommended, you can use a Personal Access Token (classic) to authenticate the runners. To do so, create a
PAT (classic) as described in the
GitHub Documentation.
Save this to the value specified by ssm_github_token_path
using the following command, adjusting the AWS profile and
region as explained above:
AWS_PROFILE=acme-core-gbl-auto-terraform AWS_REGION=us-west-2 chamber write github-action-runners github-auth-secret -- "<PAT>"
Using Runner Groups
GitHub supports grouping runners into distinct
Runner Groups,
which allow you to have different access controls for different runners. Read the linked documentation about creating
and configuring Runner Groups, which you must do through the GitHub Web UI. If you choose to create Runner Groups, you
can assign one or more Runner Sets (from the runners
map) to groups (only one group per runner set, but multiple sets
can be in the same group) by including group: <Runner Group Name>
in the runner configuration. We recommend including
it immediately after github_url
.
Interaction with Karpenter or other EKS autoscaling solutions
Kubernetes cluster autoscaling solutions generally expect that a Pod runs a service that can be terminated on one Node and restarted on another with only a short duration needed to finish processing any in-flight requests. When the cluster is resized, the cluster autoscaler will do just that. However, GitHub Action Runner Jobs do not fit this model. If a Pod is terminated in the middle of a job, the job is lost. The likelihood of this happening is increased by the fact that the Action Runner Controller Autoscaler is expanding and contracting the size of the Runner Pool on a regular basis, causing the cluster autoscaler to more frequently want to scale up or scale down the EKS cluster, and, consequently, to move Pods around.
To handle these kinds of situations, Karpenter respects an annotation on the Pod:
spec:
template:
metadata:
annotations:
karpenter.sh/do-not-evict: "true"
When you set this annotation on the Pod, Karpenter will not voluntarily evict it. This means that the Pod will stay on the Node it is on, and the Node it is on will not be considered for deprovisioning (scale down). This is good because it means that the Pod will not be terminated in the middle of a job. However, it also means that the Node the Pod is on will remain running until the Pod is terminated, even if the node is underutilized and Karpenter would like to get rid of it.
Since the Runner Pods terminate at the end of the job, this is not a problem for the Pods actually running jobs.
However, if you have set minReplicas > 0
, then you have some Pods that are just idling, waiting for jobs to be
assigned to them. These Pods are exactly the kind of Pods you want terminated and moved when the cluster is
underutilized. Therefore, when you set minReplicas > 0
, you should NOT set karpenter.sh/do-not-evict: "true"
on
the Pod.
Updating CRDs
When updating the chart or application version of gha-runner-scale-set-controller
, it is possible you will need to
install new CRDs. Such a requirement should be indicated in the gha-runner-scale-set-controller
release notes and may
require some adjustment to this component.
This component uses helm
to manage the deployment, and helm
will not auto-update CRDs. If new CRDs are needed,
follow the instructions in the release notes for the Helm chart or gha-runner-scale-set-controller
itself.
Useful Reference
- Runner Scale Set Controller's Helm chart values.yaml
- Runner Scale Set's Helm chart values.yaml
- Runner Scale Set's Docker image and how to create your own
When reviewing documentation, code, issues, etc. for self-hosted GitHub action runners or the Actions Runner Controller
(ARC), keep in mind that there are 2 implementations going by that name. The original implementation, which is now
deprecated, uses the actions.summerwind.dev
API group, and is at times called the Summerwind or Legacy implementation.
It is primarily described by documentation in the
actions/actions-runner-controller GitHub repository itself.
The new implementation, which is the one this component uses, uses the actions.github.com
API group, and is at times
called the GitHub implementation or "Runner Scale Sets" implementation. The new implementation is described in the
official
GitHub documentation.
Feature requests about the new implementation are officially directed to the
Actions category of GitHub community discussion.
However, Q&A and community support is directed to the actions/actions-runner-controller
repo's
Discussion section, though beware that discussions
about the old implementation are mixed in with discussions about the new implementation.
Bug reports for the new implementation are still filed under the actions/actions-runner-controller
repo's
Issues tab, though again, these are mixed in with bug
reports for the old implementation. Look for the gha-runner-scale-set
label to find issues specific to the new
implementation.
Requirements
Name | Version |
---|---|
<a name="requirement_terraform"></a> terraform | >= 1.3.0 |
<a name="requirement_aws"></a> aws | >= 4.9.0 |
<a name="requirement_helm"></a> helm | >= 2.0 |
<a name="requirement_kubernetes"></a> kubernetes | >= 2.0, != 2.21.0 |
Providers
Name | Version |
---|---|
<a name="provider_aws"></a> aws | >= 4.9.0 |
<a name="provider_aws.ssm"></a> aws.ssm | >= 4.9.0 |
<a name="provider_kubernetes"></a> kubernetes | >= 2.0, != 2.21.0 |
Modules
Name | Source | Version |
---|---|---|
<a name="module_eks"></a> eks | cloudposse/stack-config/yaml//modules/remote-state | 1.5.0 |
<a name="module_gha_runner_controller"></a> gha_runner_controller | cloudposse/helm-release/aws | 0.10.0 |
<a name="module_gha_runners"></a> gha_runners | cloudposse/helm-release/aws | 0.10.0 |
<a name="module_iam_roles"></a> iam_roles | ../../account-map/modules/iam-roles | n/a |
<a name="module_this"></a> this | cloudposse/label/null | 0.25.0 |
Resources
Name | Type |
---|---|
kubernetes_namespace.controller | resource |
kubernetes_namespace.runner | resource |
kubernetes_secret_v1.controller_image_pull_secret | resource |
kubernetes_secret_v1.controller_ns_github_secret | resource |
kubernetes_secret_v1.github_secret | resource |
kubernetes_secret_v1.image_pull_secret | resource |
aws_eks_cluster_auth.eks | data source |
aws_ssm_parameter.github_token | data source |
aws_ssm_parameter.image_pull_secret | data source |
Inputs
Name | Description | Type | Default | Required |
---|---|---|---|---|
<a name="input_additional_tag_map"></a> additional_tag_map | Additional key-value pairs to add to each map in tags_as_list_of_maps . Not added to tags or id .<br>This is for some rare cases where resources want additional configuration of tags<br>and therefore take a list of maps with tag key, value, and additional configuration. | map(string) | {} | no |
<a name="input_attributes"></a> attributes | ID element. Additional attributes (e.g. workers or cluster ) to add to id ,<br>in the order they appear in the list. New attributes are appended to the<br>end of the list. The elements of the list are joined by the delimiter <br>and treated as a single ID element. | list(string) | [] | no |
<a name="input_charts"></a> charts | Map of Helm charts to install. Keys are "controller" and "runner_sets". | <pre>map(object({<br> chart_version = string<br> chart = optional(string, null) # defaults according to the key to "gha-runner-scale-set-controller" or "gha-runner-scale-set"<br> chart_description = optional(string, null) # visible in Helm history<br> chart_repository = optional(string, "oci://ghcr.io/actions/actions-runner-controller-charts")<br> wait = optional(bool, true)<br> atomic = optional(bool, true)<br> cleanup_on_fail = optional(bool, true)<br> timeout = optional(number, null)<br> }))</pre> | n/a | yes |
<a name="input_context"></a> context | Single object for setting entire context at once.<br>See description of individual variables for details.<br>Leave string and numeric variables as null to use default value.<br>Individual variable settings (non-null) override settings in context object,<br>except for attributes, tags, and additional_tag_map, which are merged. | any | <pre>{<br> "additional_tag_map": {},<br> "attributes": [],<br> "delimiter": null,<br> "descriptor_formats": {},<br> "enabled": true,<br> "environment": null,<br> "id_length_limit": null,<br> "label_key_case": null,<br> "label_order": [],<br> "label_value_case": null,<br> "labels_as_tags": [<br> "unset"<br> ],<br> "name": null,<br> "namespace": null,<br> "regex_replace_chars": null,<br> "stage": null,<br> "tags": {},<br> "tenant": null<br>}</pre> | no |
<a name="input_controller"></a> controller | Configuration for the controller. | <pre>object({<br> image = optional(object({<br> repository = optional(string, null)<br> tag = optional(string, null) # Defaults to the chart appVersion<br> pull_policy = optional(string, null)<br> }), null)<br> replicas = optional(number, 1)<br> kubernetes_namespace = string<br> create_namespace = optional(bool, true)<br> chart_values = optional(any, null)<br> affinity = optional(map(string), {})<br> labels = optional(map(string), {})<br> node_selector = optional(map(string), {})<br> priority_class_name = optional(string, "")<br> resources = optional(object({<br> limits = optional(object({<br> cpu = optional(string, null)<br> memory = optional(string, null)<br> }), null)<br> requests = optional(object({<br> cpu = optional(string, null)<br> memory = optional(string, null)<br> }), null)<br> }), null)<br> tolerations = optional(list(object({<br> key = string<br> operator = string<br> value = optional(string, null)<br> effect = string<br> })), [])<br> log_level = optional(string, "info")<br> log_format = optional(string, "json")<br> update_strategy = optional(string, "immediate")<br> })</pre> | n/a | yes |
<a name="input_create_github_kubernetes_secret"></a> create_github_kubernetes_secret | If true , this component will create the Kubernetes Secret that will be used to get<br>the GitHub App private key or GitHub PAT token, based on the value retrieved<br>from SSM at the var.ssm_github_secret_path . WARNING: This will cause<br>the secret to be stored in plaintext in the Terraform state.<br>If false , this component will not create a secret and you must create it<br>(with the name given by var.github_kubernetes_secret_name ) in every<br>namespace where you are deploying runners (the controller does not need it). | bool | true | no |
<a name="input_create_image_pull_kubernetes_secret"></a> create_image_pull_kubernetes_secret | If true and image_pull_secret_enabled is true , this component will create the Kubernetes image pull secret resource,<br>using the value in SSM at the path specified by ssm_image_pull_secret_path .<br>WARNING: This will cause the secret to be stored in plaintext in the Terraform state.<br>If false , this component will not create a secret and you must create it<br>(with the name given by var.github_kubernetes_secret_name ) in every<br>namespace where you are deploying controllers or runners. | bool | true | no |
<a name="input_delimiter"></a> delimiter | Delimiter to be used between ID elements.<br>Defaults to - (hyphen). Set to "" to use no delimiter at all. | string | null | no |
<a name="input_descriptor_formats"></a> descriptor_formats | Describe additional descriptors to be output in the descriptors output map.<br>Map of maps. Keys are names of descriptors. Values are maps of the form<br>{<br> format = string<br> labels = list(string)<br>} <br>(Type is any so the map values can later be enhanced to provide additional options.)<br>format is a Terraform format string to be passed to the format() function.<br>labels is a list of labels, in order, to pass to format() function.<br>Label values will be normalized before being passed to format() so they will be<br>identical to how they appear in id .<br>Default is {} (descriptors output will be empty). | any | {} | no |
<a name="input_eks_component_name"></a> eks_component_name | The name of the eks component | string | "eks/cluster" | no |
<a name="input_enabled"></a> enabled | Set to false to prevent the module from creating any resources | bool | null | no |
<a name="input_environment"></a> environment | ID element. Usually used for region e.g. 'uw2', 'us-west-2', OR role 'prod', 'staging', 'dev', 'UAT' | string | null | no |
<a name="input_github_app_id"></a> github_app_id | The ID of the GitHub App to use for the runner controller. Leave empty if using a GitHub PAT. | string | null | no |
<a name="input_github_app_installation_id"></a> github_app_installation_id | The "Installation ID" of the GitHub App to use for the runner controller. Leave empty if using a GitHub PAT. | string | null | no |
<a name="input_github_kubernetes_secret_name"></a> github_kubernetes_secret_name | Name of the Kubernetes Secret that will be used to get the GitHub App private key or GitHub PAT token. | string | "gha-github-secret" | no |
<a name="input_helm_manifest_experiment_enabled"></a> helm_manifest_experiment_enabled | Enable storing of the rendered manifest for helm_release so the full diff of what is changing can been seen in the plan | bool | false | no |
<a name="input_id_length_limit"></a> id_length_limit | Limit id to this many characters (minimum 6).<br>Set to 0 for unlimited length.<br>Set to null for keep the existing setting, which defaults to 0 .<br>Does not affect id_full . | number | null | no |
<a name="input_image_pull_kubernetes_secret_name"></a> image_pull_kubernetes_secret_name | Name of the Kubernetes Secret that will be used as the imagePullSecret. | string | "gha-image-pull-secret" | no |
<a name="input_image_pull_secret_enabled"></a> image_pull_secret_enabled | Whether to configure the controller and runners with an image pull secret. | bool | false | no |
<a name="input_kube_data_auth_enabled"></a> kube_data_auth_enabled | If true , use an aws_eks_cluster_auth data source to authenticate to the EKS cluster.<br>Disabled by kubeconfig_file_enabled or kube_exec_auth_enabled . | bool | false | no |
<a name="input_kube_exec_auth_aws_profile"></a> kube_exec_auth_aws_profile | The AWS config profile for aws eks get-token to use | string | "" | no |
<a name="input_kube_exec_auth_aws_profile_enabled"></a> kube_exec_auth_aws_profile_enabled | If true , pass kube_exec_auth_aws_profile as the profile to aws eks get-token | bool | false | no |
<a name="input_kube_exec_auth_enabled"></a> kube_exec_auth_enabled | If true , use the Kubernetes provider exec feature to execute aws eks get-token to authenticate to the EKS cluster.<br>Disabled by kubeconfig_file_enabled , overrides kube_data_auth_enabled . | bool | true | no |
<a name="input_kube_exec_auth_role_arn"></a> kube_exec_auth_role_arn | The role ARN for aws eks get-token to use | string | "" | no |
<a name="input_kube_exec_auth_role_arn_enabled"></a> kube_exec_auth_role_arn_enabled | If true , pass kube_exec_auth_role_arn as the role ARN to aws eks get-token | bool | true | no |
<a name="input_kubeconfig_context"></a> kubeconfig_context | Context to choose from the Kubernetes config file.<br>If supplied, kubeconfig_context_format will be ignored. | string | "" | no |
<a name="input_kubeconfig_context_format"></a> kubeconfig_context_format | A format string to use for creating the kubectl context name when<br>kubeconfig_file_enabled is true and kubeconfig_context is not supplied.<br>Must include a single %s which will be replaced with the cluster name. | string | "" | no |
<a name="input_kubeconfig_exec_auth_api_version"></a> kubeconfig_exec_auth_api_version | The Kubernetes API version of the credentials returned by the exec auth plugin | string | "client.authentication.k8s.io/v1beta1" | no |
<a name="input_kubeconfig_file"></a> kubeconfig_file | The Kubernetes provider config_path setting to use when kubeconfig_file_enabled is true | string | "" | no |
<a name="input_kubeconfig_file_enabled"></a> kubeconfig_file_enabled | If true , configure the Kubernetes provider with kubeconfig_file and use that kubeconfig file for authenticating to the EKS cluster | bool | false | no |
<a name="input_label_key_case"></a> label_key_case | Controls the letter case of the tags keys (label names) for tags generated by this module.<br>Does not affect keys of tags passed in via the tags input.<br>Possible values: lower , title , upper .<br>Default value: title . | string | null | no |
<a name="input_label_order"></a> label_order | The order in which the labels (ID elements) appear in the id .<br>Defaults to ["namespace", "environment", "stage", "name", "attributes"].<br>You can omit any of the 6 labels ("tenant" is the 6th), but at least one must be present. | list(string) | null | no |
<a name="input_label_value_case"></a> label_value_case | Controls the letter case of ID elements (labels) as included in id ,<br>set as tag values, and output by this module individually.<br>Does not affect values of tags passed in via the tags input.<br>Possible values: lower , title , upper and none (no transformation).<br>Set this to title and set delimiter to "" to yield Pascal Case IDs.<br>Default value: lower . | string | null | no |
<a name="input_labels_as_tags"></a> labels_as_tags | Set of labels (ID elements) to include as tags in the tags output.<br>Default is to include all labels.<br>Tags with empty values will not be included in the tags output.<br>Set to [] to suppress all generated tags.<br>Notes:<br> The value of the name tag, if included, will be the id , not the name .<br> Unlike other null-label inputs, the initial setting of labels_as_tags cannot be<br> changed in later chained modules. Attempts to change it will be silently ignored. | set(string) | <pre>[<br> "default"<br>]</pre> | no |
<a name="input_name"></a> name | ID element. Usually the component or solution name, e.g. 'app' or 'jenkins'.<br>This is the only ID element not also included as a tag .<br>The "name" tag is set to the full id string. There is no tag with the value of the name input. | string | null | no |
<a name="input_namespace"></a> namespace | ID element. Usually an abbreviation of your organization name, e.g. 'eg' or 'cp', to help ensure generated IDs are globally unique | string | null | no |
<a name="input_regex_replace_chars"></a> regex_replace_chars | Terraform regular expression (regex) string.<br>Characters matching the regex will be removed from the ID elements.<br>If not set, "/[^a-zA-Z0-9-]/" is used to remove all characters other than hyphens, letters and digits. | string | null | no |
<a name="input_region"></a> region | AWS Region. | string | n/a | yes |
<a name="input_runners"></a> runners | Map of Runner Scale Set configurations, with the key being the name of the runner set.<br>Please note that the name must be in kebab-case (no underscores).<br><br>For example:<pre>hcl<br>organization-runner = {<br> # Specify the scope (organization or repository) and the target<br> # of the runner via the github_url input.<br> # ex: https://github.com/myorg/myrepo or https://github.com/myorg<br> github_url = https://github.com/myorg<br> group = "core-automation" # Optional. Assigns the runners to a runner group, for access control.<br> min_replicas = 1<br> max_replicas = 5<br>}</pre> | <pre>map(object({<br> # we allow a runner to be disabled because Atmos cannot delete an inherited map object<br> enabled = optional(bool, true)<br> github_url = string<br> group = optional(string, null)<br> kubernetes_namespace = optional(string, null) # defaults to the controller's namespace<br> create_namespace = optional(bool, true)<br> image = optional(string, "ghcr.io/actions/actions-runner:latest") # repo and tag<br> mode = optional(string, "dind") # Optional. Can be "dind" or "kubernetes".<br> pod_labels = optional(map(string), {})<br> pod_annotations = optional(map(string), {})<br> affinity = optional(map(string), {})<br> node_selector = optional(map(string), {})<br> tolerations = optional(list(object({<br> key = string<br> operator = string<br> value = optional(string, null)<br> effect = string<br> # tolerationSeconds is not supported, because Terraform requires all objects in a list to have the same keys,<br> # but tolerationSeconds must be omitted to get the default behavior of "tolerate forever".<br> # If really needed, could use a default value of 1,000,000,000 (one billion seconds = about 32 years).<br> })), [])<br> min_replicas = number<br> max_replicas = number<br><br> # ephemeral_pvc_storage and _class are ignored for "dind" mode but required for "kubernetes" mode<br> ephemeral_pvc_storage = optional(string, null) # ex: 10Gi<br> ephemeral_pvc_storage_class = optional(string, null)<br><br> kubernetes_mode_service_account_annotations = optional(map(string), {})<br><br> resources = optional(object({<br> limits = optional(object({<br> cpu = optional(string, null)<br> memory = optional(string, null)<br> ephemeral-storage = optional(string, null)<br> }), null)<br> requests = optional(object({<br> cpu = optional(string, null)<br> memory = optional(string, null)<br> ephemeral-storage = optional(string, null)<br> }), null)<br> }), null)<br> }))</pre> | {} | no |
<a name="input_ssm_github_secret_path"></a> ssm_github_secret_path | The path in SSM to the GitHub app private key file contents or GitHub PAT token. | string | "/github-action-runners/github-auth-secret" | no |
<a name="input_ssm_image_pull_secret_path"></a> ssm_image_pull_secret_path | SSM path to the base64 encoded dockercfg image pull secret. | string | "/github-action-runners/image-pull-secrets" | no |
<a name="input_ssm_region"></a> ssm_region | AWS Region where SSM secrets are stored. Defaults to var.region . | string | null | no |
<a name="input_stage"></a> stage | ID element. Usually used to indicate role, e.g. 'prod', 'staging', 'source', 'build', 'test', 'deploy', 'release' | string | null | no |
<a name="input_tags"></a> tags | Additional tags (e.g. {'BusinessUnit': 'XYZ'} ).<br>Neither the tag keys nor the tag values will be modified by this module. | map(string) | {} | no |
<a name="input_tenant"></a> tenant | ID element _(Rarely used, not included by default)_. A customer identifier, indicating who this instance of a resource is for | string | null | no |
Outputs
Name | Description |
---|---|
<a name="output_metadata"></a> metadata | Block status of the deployed release |
<a name="output_runners"></a> runners | Human-readable summary of the deployed runners |
References
- cloudposse/terraform-aws-components - Cloud Posse's upstream component
- alb-controller - Helm Chart
- alb-controller - AWS Load Balancer Controller
- actions-runner-controller Webhook Driven Scaling
- actions-runner-controller Chart Values
- How to set service account for workers spawned in Kubernetes mode
[!NOTE] This project is part of Cloud Posse's comprehensive "SweetOps" approach towards DevOps.
<details><summary><strong>Learn More</strong></summary>It's 100% Open Source and licensed under the APACHE2.
</details>
<a href="https://cloudposse.com/readme/header/link?utm_source=github&utm_medium=readme&utm_campaign=cloudposse-terraform-components/aws-eks-github-actions-runner&utm_content=readme_header_link"><img src="https://cloudposse.com/readme/header/img"/></a>
Related Projects
Check out these related projects.
- Cloud Posse Terraform Modules - Our collection of reusable Terraform modules used by our reference architectures.
- Atmos - Atmos is like docker-compose but for your infrastructure
β¨ Contributing
This project is under active development, and we encourage contributions from our community. Many thanks to our outstanding contributors:
<a href="https://github.com/cloudposse-terraform-components/aws-eks-github-actions-runner/graphs/contributors"> <img src="https://contrib.rocks/image?repo=cloudposse-terraform-components/aws-eks-github-actions-runner&max=24" /> </a>π Bug Reports & Feature Requests
Please use the issue tracker to report any bugs or file feature requests.
π» Developing
If you are interested in being a contributor and want to get involved in developing this project or help out with Cloud Posse's other projects, we would love to hear from you!
Hit us up in Slack, in the #cloudposse
channel.
In general, PRs are welcome. We follow the typical "fork-and-pull" Git workflow.
- Review our Code of Conduct and Contributor Guidelines.
- Fork the repo on GitHub
- Clone the project to your own machine
- Commit changes to your own branch
- Push your work back up to your fork
- Submit a Pull Request so that we can review your changes
NOTE: Be sure to merge the latest changes from "upstream" before making a pull request!
π Slack Community
Join our Open Source Community on Slack. It's FREE for everyone! Our "SweetOps" community is where you get to talk with others who share a similar vision for how to rollout and manage infrastructure. This is the best place to talk shop, ask questions, solicit feedback, and work together as a community to build totally sweet infrastructure.
π° Newsletter
Sign up for our newsletter and join 3,000+ DevOps engineers, CTOs, and founders who get insider access to the latest DevOps trends, so you can always stay in the know. Dropped straight into your Inbox every week β and usually a 5-minute read.
π Office Hours <a href="https://cloudposse.com/office-hours?utm_source=github&utm_medium=readme&utm_campaign=cloudposse-terraform-components/aws-eks-github-actions-runner&utm_content=office_hours"><img src="https://img.cloudposse.com/fit-in/200x200/https://cloudposse.com/wp-content/uploads/2019/08/Powered-by-Zoom.png" align="right" /></a>
Join us every Wednesday via Zoom for your weekly dose of insider DevOps trends, AWS news and Terraform insights, all sourced from our SweetOps community, plus a live Q&A that you canβt find anywhere else. It's FREE for everyone!
About
This project is maintained by <a href="https://cpco.io/homepage?utm_source=github&utm_medium=readme&utm_campaign=cloudposse-terraform-components/aws-eks-github-actions-runner&utm_content=">Cloud Posse, LLC</a>. <a href="https://cpco.io/homepage?utm_source=github&utm_medium=readme&utm_campaign=cloudposse-terraform-components/aws-eks-github-actions-runner&utm_content="><img src="https://cloudposse.com/logo-300x69.svg" align="right" /></a>
We are a DevOps Accelerator for funded startups and enterprises. Use our ready-to-go terraform architecture blueprints for AWS to get up and running quickly. We build it with you. You own everything. Your team wins. Plus, we stick around until you succeed.
<a href="https://cpco.io/commercial-support?utm_source=github&utm_medium=readme&utm_campaign=cloudposse-terraform-components/aws-eks-github-actions-runner&utm_content=commercial_support"><img alt="Learn More" src="https://img.shields.io/badge/learn%20more-success.svg?style=for-the-badge"/></a>
Your team can operate like a pro today.
Ensure that your team succeeds by using our proven process and turnkey blueprints. Plus, we stick around until you succeed.
<details> <summary>π <strong>See What's Included</strong></summary>- Reference Architecture. You'll get everything you need from the ground up built using 100% infrastructure as code.
- Deployment Strategy. You'll have a battle-tested deployment strategy using GitHub Actions that's automated and repeatable.
- Site Reliability Engineering. You'll have total visibility into your apps and microservices.
- Security Baseline. You'll have built-in governance with accountability and audit logs for all changes.
- GitOps. You'll be able to operate your infrastructure via Pull Requests.
- Training. You'll receive hands-on training so your team can operate what we build.
- Questions. You'll have a direct line of communication between our teams via a Shared Slack channel.
- Troubleshooting. You'll get help to triage when things aren't working.
- Code Reviews. You'll receive constructive feedback on Pull Requests.
- Bug Fixes. We'll rapidly work with you to fix any bugs in our projects.
<a href="https://cloudposse.com/readme/commercial-support/link?utm_source=github&utm_medium=readme&utm_campaign=cloudposse-terraform-components/aws-eks-github-actions-runner&utm_content=readme_commercial_support_link"><img src="https://cloudposse.com/readme/commercial-support/img"/></a>
License
<a href="https://opensource.org/licenses/Apache-2.0"><img src="https://img.shields.io/badge/License-Apache%202.0-blue.svg?style=for-the-badge" alt="License"></a>
<details> <summary>Preamble to the Apache License, Version 2.0</summary> <br/> <br/>Licensed to the Apache Software Foundation (ASF) under one
or more contributor license agreements. See the NOTICE file
distributed with this work for additional information
regarding copyright ownership. The ASF licenses this file
to you under the Apache License, Version 2.0 (the
"License"); you may not use this file except in compliance
with the License. You may obtain a copy of the License at
https://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing,
software distributed under the License is distributed on an
"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
KIND, either express or implied. See the License for the
specific language governing permissions and limitations
under the License.
</details>
Trademarks
All other trademarks referenced herein are the property of their respective owners.
Copyright Β© 2017-2024 Cloud Posse, LLC
<a href="https://cloudposse.com/readme/footer/link?utm_source=github&utm_medium=readme&utm_campaign=cloudposse-terraform-components/aws-eks-github-actions-runner&utm_content=readme_footer_link"><img alt="README footer" src="https://cloudposse.com/readme/footer/img"/></a>
<img alt="Beacon" width="0" src="https://ga-beacon.cloudposse.com/UA-76589703-4/cloudposse-terraform-components/aws-eks-github-actions-runner?pixel&cs=github&cm=readme&an=aws-eks-github-actions-runner"/>