Awesome
Eunomia - a GitOps Operator for Kubernetes
Project Deprication and Archival
As of March 9, 2022 this project is no longer maintained. This repository is being archived(marked as read only).
Who is Eunomia
According to Wikipedia:
Eunomia (Greek: Εὐνομία) was a minor Greek goddess of law and legislation (her name can be translated as "good order", "governance according to good laws"), as well as the spring-time goddess of green pastures (eû means "well, good" in Greek, and νόμος, nómos, means "law", while pasturelands are called nomia).
How to Pronounce it
Eunomia is pronounced: yuˈnoʊ mi ə. The English phonetic respelling is yoo-noh-mee-uh.
You can listen to it here and here.
What is GitOps
GitOps is all about turning day 2 operations into code! Not just that, it means you start thinking about day 2 on day 1. This is a dream come true for any Operations team! GitOps leverages the strength of automation and combines it with the power of git based workflows. It is a natural evolution beyond infrastructure-as-code and builds on top of DevOps best practices.
Next Generation Change Management
Especially in large Enterprises, Change Management is usually a painful experience. GitOps allows to take a lot of that pain out and streamline the process itself. It does so by still providing what the process tries to accomplish (and thus still meet audit requirements), but does so in a way that is much faster, much more secure, and much more reliable.
Your changes now all of a sudden provide:
- Version Control
- Peer Reviews
- Approvals
- Audit Trail
- Reproducibility
- Consistency
- Reliability
What's your backout plan for your change? How about simply moving back to the previous commit and getting EXACTLY what you had before?
Purpose
The Eunomia provides the ability to implement these git-based flows for any resources in Kubernetes. Eunomia does not care if you have a plain Kubernetes, a cloud based Kubernetes (like GKE), or a complete PaaS platform based on Kubernetes (like OpenShift). Eunomia also does not care how you want to structure your data, how many repos you want to use, or which templating engine is your favourite.
Eunomia can handle straight-up (static) yaml files with the complete definition or create dynamic ones based on your templating engine of choice. Eunomia already supports Helm Charts, OpenShift Templates, and Jinja2 Templates, but can easily be extended to support others.
These templates will be merged and processed with a set of environment-specific parameters to get a list of resource manifests. Then these manifest can be created/updated/deleted in Kubernetes.
Getting started
If you want to deploy Eunomia from scratch to a local Minikube cluster, begin with the Getting started document. It will guide your through the setup step-by-step.
Vision
While this controller can certainly be used to directly populate an individual namespace with a configuration stored in git, the vision is that a hierarchy of controllers will be used to populate multiple namespaces. Ideally this approach will be used to bring a newly created cluster to a desired configured state. Only the initial seeding CR (Custom Resource) should have cluster-level permissions. Any sub-CRs should have progressively less access assigned to their service accounts.
Here is a very simple example of how the hierarchy might look like:
The main sections of the GitOpsConfig CRD (Custom Resource Definition) are described below.
Example
The configuration is described in the GitOpsConfig CRD, here is an example:
apiVersion: eunomia.kohls.io/v1alpha1
kind: GitOpsConfig
metadata:
name: simple-test
spec:
# Add fields here
templateProcessorArgs: "-e cluster_name=my_cluster_name"
templateSource:
uri: https://github.com/KohlsTechnology/eunomia
ref: master
contextDir: simple/templates
secretRef: template-gitconfig
parameterSource:
contextDir: simple/parameters
secretRef: parameter-gitconfig
triggers:
- type: Change
- type: Webhook
- type: Periodic
cron: "0 * * * *"
serviceAccountRef: "mysvcaccount"
templateProcessorImage: "quay.io/kohlstechnology/eunomia-base:latest"
resourceDeletionMode: "Delete"
resourceHandlingMode: "Apply"
TemplateSource and ParameterSource and TemplateProcessorArgs
The TemplateSource
and ParameterSource
specify where the templates and the parameters are stored. The exact contents of these locations depend on the templating engine that has been selected.
The TemplateProcessorArgs
can be used to pass arguments/flags to the template processor. They can be accessed by the template processor in an environment variable named TEMPLATE_PROCESSOR_ARGS
.
The fields of this section are:
templateProcessorArgs: "-e cluster_name=my_cluster_name"
templateSource:
uri: https://github.com/KohlsTechnology/eunomia
ref: master
contextDir: simple/templates
httpProxy: <http proxy>
httpsProxy: <https proxy>
noProxy: <no Proxy>
SecretRef: <gitconfig and credentials secret>
parameterSource:
uri: https://github.com/KohlsTechnology/eunomia
ref: master
contextDir: seed/parameters
httpProxy: <http proxy>
httpsProxy: <https proxy>
noProxy: <no Proxy>
SecretRef: <gitconfig and credentials secret>
These are the mandatory constraints and default behaviors of the fields:
field name | mandatory | default |
---|---|---|
uri | yes | N/A |
ref | no | master |
contextDir | no | . |
httpProxy | no | |
httpsProxy | no | |
noProxy | no | |
SecretRef | no |
If a secret is provided, then it is assumed that the connection to Git requires authentication. See the [Git Authentication](#git\ authentication) section below for more details.
If the uri
is not specified in the parameterSource
section, then it will default to the uri
specified under templateSource
.
Git Submodules
Some helm charts might require the configuration to be part of the chart itself (you can't read files from outside the chart). Loading files into a configmap is one example of this.
Separating the charts (templateSource) from the actual configuration (parameterSource) is a best practice. This allows you to separate your code (templates) from your configuration, which helps tremendously with change management.
One way to go about this is to use the config repo as a submodule and point to the master branch. During development you can of course point against another branch, just make sure you correct it in .gitmodules
before the PR gets merged.
Add submodule to track master branch
git submodule add -b master <repo-url>
Checking out a repo with submodules
git clone <repo-url>
cd <repo>
git submodule init
git submodule update --recursive --remote
parameterSource processing
Eunomia uses the yq command to merge all yaml files in the specified folder. You have to be careful, if you have the same variable name in multiple files. Dictionaries will merge, lists will get overwritten.
Variable Hierarchy
You can provide a file hierarchy.lst
, to allow a variable hierarchy. This will allow you to specify a default value and overwrite it on an environment level if necessary. This will greatly simplify your configuration and allows for deduplication of data, making your operational life a lot easier.
The contents of the file are simply relative path names, with the base being contextDir
.
Example hierarchy.lst
:
../defaults #this is first ... lowest priority
../marketing #this is second
../development #this is third ... highest priority
In this case it will load all yaml files from ../defaults
, then merge it with everything in ../marketing
, and lastly merges it with everything in ../development
.
You can also use the relative path ./
, which means it'll also load the variables defined in contextDir
directly (same folder that as hierarchy.lst
). You can insert ./
in whatever order you want in the hierarchy.lst
- it will determine its priority.
Upcoming features
Once issue #4 is resolved, you will be able to specify variable names to dynamically determine the correct folder. This will allow you to only have one hierarchy.lst
. (Technically, it is actually already possible to use environment variables, but without #4, there are just none set that would be of any practical use in hierarchy.lst.)
Git Authentication
Specifying a SecretRef
will automatically turn on git authentication. The secrets for the template and parameter repos will be mounted respectively in the /template-gitconfig
and /parameter-gitconfig
of the job pod.
The referenced secrets must be available and how they are provisioned is beyond the scope of this operator. See the Vision paragraph on how to build a hierarchical structure, where the resources needed to run a given GitOpsConfig are configured by a predecessor GitpOpsConfig instance.
This secret will be linked from ~/
of the used running the pod. The secret must contain a .gitconfig
file and may contain other files. The passed .gitconfig
will be used during the git operations. It is advised to reference any additional files via the absolute path.
Username and password authentication
For username and password based authentication create the following .gitconfig
:
[http]
sslCAInfo = /template-gitconfig/ca.crt
[user]
name = gitconfig
[credential]
helper = store
When the credential helper is of type store
, credentials are by default retrieved from the ~/.git-credentials
file. This file should also be added to the secret and has the following format:
https://<username>:<password>@<git server fqdn>
Don't forget to provide the ca.crt
file to the secret.
Certificate based authentication
For certificate based authentication, create the following .gitconfig
:
[core]
sshCommand = "ssh -i /template-gitconfig/mykey.rsa -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no"
and add the mykey.rsa
file to the secret.
Job templates
For Eunomia to work properly there is a need for a specific Job
or a CronJob
.
A Job default template
and a CronJob default template
are built into the Dockerfile.
If you want to provide your own job templates, set the env variables JOB_TEMPLATE
and CRONJOB_TEMPLATE
. Their values should be set to paths, where appropriate yaml files can be found.
The files themselves have to be accessible in the pod. To achieve this, you can for instance add ConfigMap data to a Volume
.
Triggers
You can enable one or multiple triggers.
Name | Description |
---|---|
Change | This triggers every time the CR is changed, including when it is created. |
Periodic | Periodically apply the configuration. This can be used to either schedule changes for a specific time, use it for drift management to revert any changes, or as a safeguard in case webhooks were missed. It uses a cron-style expression. |
Webhook | This triggers when something on git changes. You have to configure the webhook yourself. For branches use just branch name in GitOpsConfig CR ref , but if you want webhook working for git tag, use refs/tags/[tag_name]. |
GitHub webhook configuration
To set up GitHub webhook follow this GitHub documentation.
Create route on port 8080 to eunomia-service and use this route as GitHub webhook Payload URL
with added webhook/ endpoint at the end.
Content type needs to be set to application/json
.
Choose Just the push event
to trigger webhook.
Template Engine
When it's time to apply a configuration, the GitOps controller runs a job pod. The image of the job pod can be specified in the templateProcessorImage
field.
This is the plugin mechanism to support multiple template engines.
A base image is provided that can be inherited to simplify the process of adding support for a new templating engine.
The base image provides the following workflow:
-
gitClone.sh
: This will clone the template and parameter repos. It is expected that there will be no need to customize this. Any required changes are most likely worthy of a pull-request upstream. -
discoverEnvironment.sh
: This will create a set of environment variables that are specific to the target Kubernetes environment. Currently the following variables are supported:Name Description CA_BUNDLE
Path to the platform-level CA bundle SERVICE_CA_BUNDLE
Path to the service-level CA bundle NAMESPACE
Current namespace -
hierarchy
: processes all the parameter files and generates/tmp/eunomia_values_processed.yaml
. It currently supports the following features:- Merging of all existing yaml files in the
CLONED_PARAMETER_GIT_DIR
location, into a single file for processing by the templating engine. - Substitution of variables with environment variables within the hierarchy.
- Substitution of variables with environment variables within the final yaml file.
- Merging of all existing yaml files in the
-
processTemplate.sh
: This file needs to be overwritten in order to support a different templating engine. The contract is the following:- Templates are available at the location specified by the variable:
CLONED_TEMPLATE_GIT_DIR
- Parameters are available at the location specified by the variable:
CLONED_PARAMETER_GIT_DIR
- After the template processing completes, the processed manifests should be stored at the location of this variable:
MANIFEST_DIR
- Templates are available at the location specified by the variable:
-
resourceManager.sh
: Processes the resources inMANIFEST_DIR
. One or more files can be present, and all will be processed.
Currently the following templating engines are supported (follow the link to see examples of how new template processors can be added):
serviceAccountRef
This is the service account used by the job pod that will process the resources. The service account must be present in the same namespace as the one where the GitOpsConfig CR is and must have enough permission to manage the resources. It is out of scope of this controller how that service account is provisioned, although you can use a different GitOpsConfig CR to provision it (seeding CR).
In the Helm deployment, the ClusterRole
eunomia-cluster-list
provides list
access to all resources, and will be provisioned if you set .Values.eunomia.operator.deployment.clusterViewer
to true
.
This ClusterRole
is intended to be used in a ClusterRoleBinding
with "job runner" service accounts so they can find all of the resources that it owns.
Without the ClusterRoleBinding
, the jobs can still successfully run, however there will be error logs stating it can not find any cluster scoped resources.
Resource Handling Mode
This field specifies how resources should be handled, once the templates are processed. The following modes are currently supported:
Apply
, which is roughly equivalent tokubectl apply
. Additionally, auto-detection of resources removed from git is performed, and they're deleted from the cluster. This is done by marking all the resources with a custom label, and removing resources for which the label was not touched bykubectl apply
.Patch
. Patch requires objects to already exists and will patch them. It's useful when customizing objects that are provided through other means.Create
, equivalent tokubectl create
. Template processors which take over the resource handling phase are not required to support this mode.Replace
, equivalent tokubectl replace
. Template processors which take over the resource handling phase are not required to support this mode.Delete
, equivalent tokubectl delete
. Template processors which take over the resource handling phase are not required to support this mode.None
. In some cases there may be template processors or automation frameworks where the processing of templates and handling of generated resources are a single step. In that case, Eunomia can be configured to skip the built-in resource handling step.
Resource Deletion Mode
This field specifies how to handle resources when the GitOpsConfig object is deleted. Two options are available:
Retain
, resources previously created are left intact.Delete
, resources are delete with thecascade
option.None
, resource deletion is not handled at all.
Installing Eunomia
Installing on Kubernetes Using Helm
See the helm README for details.
Installing on Kubernetes Using Operator Hub
See the getting started document for details.
Examples / Demos
We've created several examples for you to test out Eunomia. See EXAMPLES for details.
Monitoring
Monitoring with Prometheus
Prometheus is an open-source systems monitoring and alerting toolkit.
Prometheus collects metrics from monitored targets by scraping metrics HTTP endpoints.
-
scrape_configs
controls what resources Prometheus monitors. -
kubernetes_sd_configs
Kubernetes SD configurations allow retrieving scrape targets. Please see kubernetes_sd_configs for details. -
Additionally,
relabel_configs
allow advanced modifications to any target and its labels before scraping.
By default, the metrics in Operator SDK are exposed on 0.0.0.0:8383/metrics
For more information, see Metrics in Operator SDK
Usage:
scrape_configs:
- job_name: 'kubernetes-service-endpoints'
kubernetes_sd_configs:
- role: endpoints
relabel_configs:
- source_labels: [__meta_kubernetes_namespace]
action: keep
regex: test-eunomia-operator
You can find additional examples on their GitHub page.
Verify metrics port:
kubectl exec POD-NAME
curl localhost:8383/metrics -n NAMESPACE
(e.g. kubectl exec eunomia-operator-5b9b664cfc-6rdrh curl localhost:8383/metrics -n test-eunomia-operator
)
Kubernetes Events
Eunomia emits the following events in the namespace of the GitOpsConfig CR:
- JobSuccessful - when a Job applying the CR finished successfully
- JobFailed - when a Job applying the CR has finished with a failure (after all retries have failed)
Development
Please see our development documentation for details.
Troubleshooting
Please see our troubleshooting guide for details.
License
See LICENSE for details.
Code of Conduct
See CODE_OF_CONDUCT.md for details.