NOTE: This project is effectively a fork of a mini-lab of DockerCon 2016 US Orchestration Lab. Just pulled it into a standalone project for easier building.
Orchestration with Swarm Computing
Time: Approximately 40 minutes
In this lab you will deploy a Dockerized application to a single host and test the application. You will then configure Docker for Swarm Computing and deploy the same app across multiple hosts. You will then see how to scale the application and move the workload across different hosts
You will complete the following steps in this lab:
- Deploy a single host application
- Configure Docker for Swarm Computing
- Deploy the application across multiple hosts
- Scale the application
- Drain a node and reschedule the containers
You will need all of the following in order to complete this lab:
- Three nodes running Docker v1.12.x. Each VM will be referenced as v112node0, v112node1, and v112node2.
- A Docker ID. Creating a Docker ID is free, and allows you to push and pull images from Docker Hub. This link describes how to create a Docker ID (you only need to complete the procedure up to step 2.3).
<a name="deploy-application"></a>Step 1: Deploy a single host application
If you would like to skip using a Dockerfile and jump straight deploying on a cluster then go straight to Step 2.0.
In this step you will deploy a simple application that runs on a single Docker host. In order to do that, you will complete the following:
- Clone the app's GitHub repo
- Dockerize the app
- Run the app
- Push the image to Docker Hub
The application you will deploy is the
cats application. It is a simple 1-container application that displays random pictures of cats, because why not!! It is a flask application written in Python that pulls the images from public URLs.
Step 1.1 - Log into VMs and verify that Docker is running
SSH to your v112node0 with the username of
labuser. You should have the hostnames in an email titled "Docker Labs VMs Ready"
The command to SSH into v112node0 will look something like the following:
You will be asked for a password. The password is in the email with the DNS names of the lab VMs.
An example of the sign in ...
ssh firstname.lastname@example.org The authenticity of host 'v112node0-9128f1906df54acda5044f56a1a86b07-2.cloudapp.net (184.108.40.206)' can't be established. ECDSA key fingerprint is SHA256:3L9UZQTSCOUkzVuyO5z3H7jLedSp8/5AquUtSZ8ydEE. Are you sure you want to continue connecting (yes/no)? yes Warning: Permanently added 'v112node0-9128f1906df54acda5044f56a1a86b07-2.cloudapp.net,220.127.116.11' (ECDSA) to the list of known hosts. email@example.com's password: Welcome to Ubuntu 14.04.4 LTS (GNU/Linux 4.2.0-23-generic x86_64)
Verify that Docker is running. This lab is based on a pre-release version of Docker Engine 1.12 so that we can show you the latest Swarm features. We are running version
1.12.0-rc1on these VMs.
labuser@v112node0:~$ docker version Client: Version: 1.12.0-rc1 API version: 1.24 Go version: go1.6.2 Git commit: 1f136c1 Built: Wed Jun 15 05:16:17 2016 OS/Arch: linux/amd64 Server: Version: 1.12.0-rc1 API version: 1.24 Go version: go1.6.2 Git commit: 1f136c1 Built: Wed Jun 15 05:16:17 2016 OS/Arch: linux/amd64
We are going to copy the contents of the application code to v112node0. Use
gitto clone the
```bash labuser@v112node0:~$ git clone https://github.com/mark-church/cats.git
Cloning into 'cats'...
remote: Counting objects: 10, done.
remote: Compressing objects: 100% (6/6), done.
remote: Total 10 (delta 0), reused 7 (delta 0), pack-reused 0
Unpacking objects: 100% (10/10), done.
Checking connectivity... done.
``` The repo contains all of the files and code that is required to create a container image of the application. The command above copies (clones) the repo into a new directory on your machine called `cats`.
Change directory to
catsand examine the list of files with 'tree'.
labuser@v112node0:~/$ cd cats labuser@v112node0:~/cats$ tree . ├── app.py ├── Dockerfile ├── images │ ├── browser.png │ ├── cat.png │ └── food.png ├── README.md └── templates └── index.html 2 directories, 7 files
Some of the files worth knowing include the following:
Dockerfile: This file contains the recipe for the
app.py: This is the main Python module that is our Flask application.
/app/templates: This folder includes the HTML web page that is called by the Python web server.
Let's focus on the Dockerfile for a second. A Dockerfile is a text file that contains all the instructions required to build an application or service into a Docker image. This includes instructions to install packages, copy data, insert metadata, and anything else that should be included as part of the image. The Docker Engine uses Dockerfiles to create new images.
Step 1.2 - Dockerize the app
You Dockerize an application by describing it in a Dockerfile and using that Dockerfile to create a Docker image.
The following procedure will guide you through Dockerizing the the
Make sure you're logged in to v112node0 and in the
Inspect the contents of the Dockerfile.
ubuntu@v112node0:~/cats$ cat Dockerfile FROM ubuntu:14.04 RUN sudo apt-get update && apt-get -y install python-pip RUN sudo pip install flask==0.10.1 COPY . /usr/bin WORKDIR /usr/bin EXPOSE 5000 CMD ["python", "./app.py"]
Let's have a quick look at the contents of the Dockerfile.
FROM: The FROM instruction sets the base image that all subsequent instructions will build on top of. The base image can be any valid Docker image including other peoples' images. In this exercise you are starting with the
ubuntu:14.04image as your base image.
RUN: The RUN instruction executes commands while the image is being built. Each RUN instruction creates a new image layer. The RUN instructions in this Dockerfile are updating the local
aptpackage lists from source, and installing some packages that the Flask framework requires.
ADD: The ADD instruction copies files or directories from the Docker host and adds them to the filesystem of the container. In this particular Dockerfile the
appdirectory that was just cloned to v112node0 form GitHub is copied to the image at
EXPOSE: The EXPOSE instruction lists the ports that the container will expose at runtime. This image will expose port 5000.
CMD: The main purpose of CMD instruction is to provide defaults for an executing container. This CMD instruction will run
./app.pyas the argument.
If you have not already signed up for a Docker Hub account please do so. Instructions are detailed in the preqresuites for this lab. Use the following
docker buildcommand to build an image from the Dockerfile.
Be sure to substitute your own Docker ID in this example
labuser@v112node0:~/cats$ docker build -t <your Docker ID>/cats .
It will take between one and three minutes to build the image. This is because the
ubuntu:14.04image has to be pulled locally and several packages have to be installed into the image.
-tflag lets you tag the image. The
<your-docker-id>/catswill be the image tag - Be sure to substitute your Docker ID. Tagging the image lets you easily identify it as well as push to Docker Hub or other Docker Registries.
The trailing period
.tells the command to send your current working directory to the Docker daemon as the build context. The build context is a fancy way of saying the Dockerfile and other files required to build the image.
The output of the
docker buildcommand will look similar to the following (many lines have been removed for brevity).
Warning: You will see text roll up the screen as packages are installed.
``` labuser@v112node0:~/cats$ docker build -t markchurch/cats .
Sending build context to Docker daemon 65.54 kB
Step 1 : FROM ubuntu:14.04
Step 2 : RUN sudo apt-get update && apt-get -y install python-pip
---> Using cache
Step 3 : RUN sudo pip install flask==0.10.1
---> Using cache
Step 4 : COPY . /usr/bin
---> Using cache
Step 5 : WORKDIR /usr/bin
---> Using cache
Step 6 : EXPOSE 5000
---> Using cache
Step 7 : CMD python ./app.py
---> Using cache
Successfully built 0968ec22c14f
Your output will be a lot more verbose than this as it includes output from
pip. Each line in the Dockerfile is creating a new container, installing the packages inside that container, and then commiting that container to a new image. The final step gives us the final image for this application.
docker imagescommand to confirm that the image is listed.
labuser@v112node0:~/cats$ docker images REPOSITORY TAG IMAGE ID CREATED SIZE markchurch/cats latest 6991f312ee28 7 seconds ago 363.5 MB ubuntu 14.04 8f1bd21bd25c 3 weeks ago 188 MB
The output of your command will show your Docker ID and not markchurch. Since we doewnloaded the
ubuntu:14.04image as an intermediary step we can see that that image has been downloaded too.
Step 1.3 - Run the App
Now that you've Dockerized the the
cats app let's go ahead and run it.
Perform all of these steps from v112node0.
docker runcommand to deploy a container with the
<your Docker ID>/catsimage.
labuser@v112node0:~/cats$ docker run -d --name cats-app -p 8000:5000 markchurch/cats 0374e0c107b4e84d4201415d1ad5dd036787ea317c8baa2062615f1f9e0b334e
This will create a container with the
catsimage. It will run in the background because of the
-dflag. We are exposing the container on port
8000of the host and mapping this to port
5000inside the container. The container ID of the container is given as a message confirming that the container has been created from our image.
Check that the container is running with
``` labuser@v112node0:~/cats$ docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
0374e0c107b4 markchurch/cats "python ./app.py" 5 minutes ago Up 5 minutes 0.0.0.0:8000->5000/tcp cats-app
This output shows us the container ID, the image used to build the container, the status of the container, and also the port-mapping between the container interface and the host interface. This confirms that our container is running.
Point your web browser to the app.
Paste the public DNS or public IP of v112node0 into your web browser with port
<img src="images/browser.png" width=100%>
You should see the following ...
<img src="images/cat.png" width=400px>
Note that the container ID on the `cats-app` web page should match the container ID of the `docker ps` command. Try refreshing the browser a couple times. You will cycle through some fantastics cats!
Step 1.4 - Push the image to Docker Hub
Now that we have verified that our application works we will push the newly created image to Docker Hub so that you can pull it form other nodes.
Perform the following steps from v112node0.
Login with your Docker ID.
labuser@v112node0:~/cats$ docker login Login with your Docker ID to push and pull images from Docker Hub. If you don't have a Docker ID, head over to https://hub.docker.com to create one. Username: <your-docker-id> Password: Login Succeeded
Push your image to Docker Hub. Later in the lab you will be pulling this image to different hosts.
Remember to substitute your Docker ID.
```bash labuser@v112node0:~/cats$ docker push markchurch/cats
The push refers to a repository [docker.io/markchurch/cats]
5f70bf18a086: Mounted from library/ubuntu
6f8be37bd578: Mounted from library/ubuntu
9f7ab087e6e6: Mounted from library/ubuntu
dc109d4b4ccf: Mounted from library/ubuntu
a7e1c363defb: Mounted from library/ubuntu
latest: digest: sha256:5fdee62784552531475c02afff4cdcc12f95a050dbdd4bce53da7ee3d589d582 size: 1992
This will push your newly built image to your own public repository on Docker Hub called
<your-docker-id>/cats. The Docker Hub is the default Docker Registry and is hosted on the public internet. You can also push images to Docker Trusted Registry as well as third party registries.
If you go to
hub.docker.comand login then you will be able to see the repository and image tag that you just created.
<img src="images/hub.png" width=600px>
Before proceeding to the next task, clear all of the containers on v112node0 by running the following command.
labuser@v112node0:~/cats$ docker rm -f $(docker ps -q) 385f1c7a587c
Verify that the command worked and there are no running or stopped containers on the host.
labuser@v112node0:~/cats$ docker ps CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES labuser@v112node0:~/cats$
Now if you refresh your web browser with your application you will see that it is no longer running.
<a name="start-cluster"></a>Step 2: Configure Swarm Mode
So far you have deployed an application to a single Docker host (node). However, real-world applications are typically deployed across multiple hosts. This improves application performance and availability, as well as allowing individual application components to scale independently. Docker has powerful native tools to help you do this.
In this step you will configure Swarm Mode. This is a new optional mode in which multiple Docker Engines form into a self-orchestrating group of engines called a swarm. Swarm mode enables new features such as services and bundles that help you deploy and manage multi-container apps across multiple Docker hosts.
You will complete the following:
- Configure Swarm mode
- Run the app
- Scale the app
- Drain nodes for maintenance and reschedule containers
For the remainder of this lab we will refer to Docker native clustering as Swarm mode. The collection of Docker engines configured for Swarm mode will be referred to as the swarm.
A swarm comprises one or more Manager Nodes and one or more Worker Nodes. The manager nodes maintain the state of swarm and schedule appication containers. The worker nodes run the application containers. As of Docker 1.12, no external backend, or 3rd party components, are required for a fully functioning swarm - everything is built-in!
In this part of the demo you will use all three of the nodes in your lab. v112node0 will be the Swarm manager, while v112node1 and v112node2 will be worker nodes. Swarm mode supports a highly available redundant manager nodes, but for the purposes of this lab you will only deploy a single manager node.
If you are just joining us on this step then feel free to use the
markchurch/cats image from the Docker Hub for the rest of the lab
Step 2.1 - Create a Manager node
If you haven't already done so, SSH in to v112node0.
For example (remember to substitute your SSH key and v112node0 for your lab):
$ ssh firstname.lastname@example.org
Get the internal/private IP address of v112node0.
labuser@v112node0:~/$ ip a ls dev eth0 | sed -n "s,.*inet *\([^/]*\)/.*,\1,p" 10.0.0.11
Make a note of this IP address as you will use it later when adding worker nodes to the swarm.
Create a Manager node on v112node0 using its internal IP address and `docker swarm init --listen-addr <IP address>:<port>. This is the address the the swarm manager is advertising itself on to other nodes wishing to joing the swarm.
Remember to substitute the IP address for the private IP of your v112node0.
labuser@v112node0:~/$ docker swarm init --listen-addr 10.0.0.11:2377 Swarm initialized: current node (0gohc21qtm7sp) is now a manager.
docker swarm initis a new command in Docker 1.12. It is entirely optional, but is all that is needed to initialize a new swarm (native Docker cluster).
2377is recommended but not mandatory. You can use a different port of your choosing.
docker infocommand to verify that v112node0 was successfully configured as a swarm manager node.
ubuntu@v112node0:~/FoodTrucks/$ docker info Containers: 0 Running: 0 <Snip> Swarm: NodeID: 3vatl908ksmjk IsManager: YES <Snip>
The swarm is now initialized with v112node0 as the only Manager node. In the next section you will add v112node1 and v112node2 as Worker nodes.
Step 2.2 - Join Worker nodes to the Swarm
You will perform the following procedure on v112node1 and v112node2. Towards the end of the procedure you will switch back to v112node0.
- Open a new SSH session to v112node1 (Keep your SSH session to v112node0 open in another tab or window).
$ ssh email@example.com
Join the swarm using the internal IP of v112node0 and the port specified when you created the Swarm.
The format of the command is as follows:
docker swarm join <internal-ip-of-manager-node>:<port>
The example below joins a new worker node to an existing swarm with a manager node with IP address of
Be sure to substitute the internal IP address of your v112node0 that you made a note of earlier.
labuser@v112node1:~/$ docker swarm join 10.0.0.11:2377 This node joined a Swarm as a worker.
Repeat steps 1 and 2 for v112node2.
Switch to v112node0 and verify that the Swarm is up and running with v112node0 as the Manager node and v112node1 and v112node2 both as Worker nodes.
Run this command on v112node0 (the Manager node).
labuser@v112node0:~/$ docker node ls ID NAME MEMBERSHIP STATUS AVAILABILITY MANAGER STATUS LEADER 5zkcnyq6uqxv4sfpzsyvs8x3s v112node1-9128f1906df54acda5044f56a1a86b07-2 Accepted Ready Active acu262y2ight8uyn6fg8yvust * v112node0-9128f1906df54acda5044f56a1a86b07-2 Accepted Ready Active Reachable Yes brr8ri58aa1ed9051drszoyh0 v112node2-9128f1906df54acda5044f56a1a86b07-2 Accepted Ready Active
docker node lscommand shows you all of the nodes that are in the swarm as well as their roles in the swarm. The
*identifies the node that you are issuing the command from.
Congratulations. You have configured a swarm with one manager node and two worker nodes.
<a name="multi-application"></a>Step 3: Deploy a Multi-Host Application
Now that you have a swarm up and running, it is time to deploy the app to it. To do this you will complete the following actions:
- Create a new overlay network for the application
- Deploy the application components as Docker services
You will perform the following procedure from v112node0.
Step 3.1 - Create a new overlay network for the application
Create a new overlay network called
catnetwill be the overlay network that our application containers live on.
labuser@v112node0:~/$ docker network create -d overlay catnet 81pjggkd57fbbl7zmheasc6m8
The network ID will be output to signal the succesful creation of a Docker network. The
-dflag let's you specify which network driver to use. In this example you are telling Docker to create a new network using the overlay driver. In Swarm Mode the overlay network does not require an external key-value store. It is integrated into the engine. The overlay provides reachability between the hosts across the underlay network. Out of the box containers on the same overlay network will be able to ping eachother without any other special configuration.
Confirm that the network was created. We can see that there are several other default networks that already exist in a Docker host.
labuser@v112node0:~/$ docker network ls NETWORK ID NAME DRIVER SCOPE 09b78531f1d5 bridge bridge local e7cf5b545252 docker_gwbridge bridge local 9311e4b9e4e1 host host local bhpf10k7w3gt ingress overlay swarm 26b729bf0d65 none null local
Now that the container network is created you are ready to deploy the application to the swarm.
Step 3.2 - Deploy the application components as Docker services
cats application is becoming very popular on the internet. People just love pictures of cats. You are going to have to scale your application to meet peak demand. You will have to do this across multiple hosts for high availability. We will use the concept of Services to scale our application easily and manage many containers as a single entity.
Services are a new concept in Docker 1.12. They work with swarms and are intended for long-running containers.
You will perform this procedure from v112node0.
catsas a Service. Remember to use your
<Docker ID>instead of
markchurch. However, if you did not complete Part 1 of this lab you are welcome to use
markchurch/catsas the application image as that will work as well.
labuser@v112node0:~/$ docker service create --name cat-app --network catnet -p 8000:5000 markchurch/cats 5qwkfpdpsm72opbxvzgzk9s5q
Verify that the
service createhas been received by the Swarm manager.
labuser@v112node0:~/$ docker service ls ID NAME SCALE IMAGE COMMAND 5qwkfpdpsm72 cat-app 1 markchurch/cats
We can insepct the progress of our individual service containers by using the
docker service taskscommand. It may take a minute or two until the
LAST STATEcolumn is in the
labuser@v112node0:~/$ docker service tasks cat-app ID NAME SERVICE IMAGE LAST STATE DESIRED STATE NODE ehfg9vq9m8bgyrxyv0abrex8c cat-app.1 cat-app markchurch/cats Running 5 minutes Running v112node0-9128f1906df54acda5044f56a1a86b07-2
The state of the service may change a couple times until it is running. The image is being downloaded from your Docker Hub repository directly to the other engines in the Swarm. Once the image is downloaded the container goes into a running state on one of the three nodes.
At this point it may not seem that we have done anything very differently than in part 1. We have again deployed a single container on a host and mapped it to port
8000. The difference here is that the container has been scheduled on a swarm cluster. The application is available on port
8000on any of the hosts in the cluster even though the
cat-appcontainer is just running on a single host. The Swarm intelligently routes routes requests to the
Check that the app is running in your browser. You can use any of the node URLs. The Swarm will advertise the published port on every host in the cluster.
Point your web browser to
http://<v112node0-public-ip>:8000. Now try pointing it to
http://<v112node1-public-ip>:8000. You should see the same
cat-appbeing served by the same container ID in both cases. Your request is being routed by the Docker engine from any Swarm node to the correct node. (The picture will change but the container ID con