Public | Automated Build

Last pushed: 2 years ago
Short Description
same as rancher/catalog-dockerfiles but updated dependencies (nginx, kopf,..)
Full Description

Kopf in docker

Tagged docker images for kopf, lmenezes/elasticsearch-kopf on docker hub.


Use docker run as you always do. You need to publish port 80
(and 443 if you use ssl) in order to have access to kopf.

Container should have access to elasticsearch. You don't
need to expose elasticsearch to end users of kopf.

It is strongly recommended to use https and basic auth
if you don't want to get hacked.

Env variables.

  • KOPF_SERVER_NAME server name for your grafana, for example
  • KOPF_ES_SERVERS elasticsearch servers in host:port[,host:port] format
  • KOPF_SSL_CERT path to ssl .crt file, enables http-to-https redirect, should be bind-mounted
  • KOPF_SSL_KEY path to ssl .key file, should be bind-mounted
  • KOPF_BASIC_AUTH_LOGIN basic auth login, if needed
  • KOPF_BASIC_AUTH_PASSWORD hashed basic auth password, if needed
  • KOPF_NGINX_INCLUDE_FILE file to include into main server of nginx (place allowed ips here)
  • KOPF_WITH_CREDENTIALS set the external setting with_credentials. Default: false
  • KOPF_THEME set the theme in external settings. Default: dark
  • KOPF_REFRESH_RATE set the external setting refresh_rate. Default: 5000


pure docker run

Running kopf with elasticsearch on,
exposing it on with ip address

docker run -d -p -e \
    -e --name kopf lmenezes/elasticsearch-kopf


An easy way to orchestrate a local docker run is fig
Install fig by fireing up pip install fig.
After create a fig file and off you go.

$ cat << EOF > fig.yml
  image: lmenezes/elasticsearch-kopf
  - 8080:80
  - KOPF_SERVER_NAME=dockerhost
$ fig up -d
Creating docker_kopf_1...

This docker container will connect to an ES instance running on the DOCKER_HOST, which exposes 9200.

Docker Pull Command
Source Repository