Awesome
Fluentd Log Collector For Kubernetes
This image captures logs from Docker containers, enriches with Kubernetes metadata & pushes to the specified Elasticsearch service deployed in your Kubernetes environment.
You can configure behaviour via the following environment variables:
ELASTICSEARCH_HOST
- the Elasticsearh host name to send data to (default:es-logging.default.svc
)ELASTICSEARCH_PORT
- the Elasticsearh port to send data to (default:9200
)ELASTICSEARCH_SCHEME
-http
orhttps
(default:http
)ELASTICSEARCH_USER
- user to connect withELASTICSEARCH_PASSWORD
- password to connect withELASTICSEARCH_DYNAMIC
- whether elasticsearch plugin configuration is dymamic, e.g. to select index name prefix based on the namespace field (default:false
)FLUENTD_FLUSH_INTERVAL
- how often to flush fluentd data (default:10s
)FLUENTD_FLUSH_THREADS
- number of threads to use to flush logs (default:1
)FLUENTD_RETRY_LIMIT
- number of retries on flush failures (default:10
)FLUENTD_DISABLE_RETRY_LIMIT
- disable retry limit (default:true
)FLUENTD_RETRY_WAIT
- time to wait between retries (default:1s
)FLUENTD_MAX_RETRY_WAIT
- maximum time to wait between retries (default:60s
)FLUENTD_BUFFER_CHUNK_LIMIT
- max buffer chunk size (default:8m
)FLUENTD_BUFFER_QUEUE_LIMIT
- max buffer queue size (default:8192
)FLUENTD_BUFFER_TYPE
- buffer type:memory
orfile
(default:memory
)FLUENTD_BUFFER_PATH
- buffer file path if using afile
buffer (default:/var/fluentd/buffer
)FLUENTD_LOGSTASH_FORMAT
- use logstash format for index names (defaulttrue
)FLUENTD_LOGSTASH_PREFIX
- the prefix to use for index names in logstash format (defaultlogstash
)KUBERNETES_PRESERVE_JSON_LOG
- whether to preserve raw JSON logs in thelog
field (default:true
)KUBERNETES_URL
- provide the URL of the Kubernetes master (default:https://kubernetes.default.svc
)KUBERNETES_VERIFY_SSL
- validate the Kubernetes SSL certificate (default:true
)
For more information on Fluentd buffer plugins & options, see http://docs.fluentd.org/articles/buffer-plugin-overview.