class: title, self-paced Container Orchestration
with Docker and Swarm
.nav[*Self-paced version*] .debug[ ``` ?? slides/mise.toml ``` These slides have been built from commit: 3ab1907 [shared/title.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/title.md)] --- class: title, in-person Container Orchestration
with Docker and Swarm
.footnote[ **Slides[:](https://www.youtube.com/watch?v=h16zyxiwDLY) https://container.training/** ] .debug[[shared/title.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/title.md)] --- ## A brief introduction - This was initially written by [Jérôme Petazzoni](https://twitter.com/jpetazzo) to support in-person, instructor-led workshops and tutorials - Over time, [multiple contributors](https://github.com/jpetazzo/container.training/graphs/contributors) also helped to improve these materials — thank you! - You can also follow along on your own, at your own pace - We included as much information as possible in these slides - We recommend having a mentor to help you ... - ... Or be comfortable spending some time reading the Docker [documentation](https://docs.docker.com/) ... - ... And looking for answers in the [Docker forums](forums.docker.com), [StackOverflow](https://stackoverflow.com/questions/tagged/docker), and other outlets .debug[[swarm/intro.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/intro.md)] --- class: self-paced ## Hands on, you shall practice - Nobody ever became a Jedi by spending their lives reading Wookiepedia - Likewise, it will take more than merely *reading* these slides to make you an expert - These slides include *tons* of demos, exercises, and examples - They assume that you have access to some Docker nodes - If you are attending a workshop or tutorial:
you will be given specific instructions to access your cluster - If you are doing this on your own:
the first chapter will give you various options to get your own cluster .debug[[swarm/intro.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/intro.md)] --- ## Accessing these slides now - We recommend that you open these slides in your browser: https://container.training/ - This is a public URL, you're welcome to share it with others! - Use arrows to move to next/previous slide (up, down, left, right, page up, page down) - Type a slide number + ENTER to go to that slide - The slide number is also visible in the URL bar (e.g. .../#123 for slide 123) .debug[[shared/about-slides.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/about-slides.md)] --- ## These slides are open source - The sources of these slides are available in a public GitHub repository: https://github.com/jpetazzo/container.training - These slides are written in Markdown - You are welcome to share, re-use, re-mix these slides - Typos? Mistakes? Questions? Feel free to hover over the bottom of the slide ... .footnote[👇 Try it! The source file will be shown and you can view it on GitHub and fork and edit it.] .debug[[shared/about-slides.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/about-slides.md)] --- ## Accessing these slides later - Slides will remain online so you can review them later if needed (let's say we'll keep them online at least 1 year, how about that?) - You can download the slides using that URL: https://container.training/slides.zip (then open the file `swarm-selfpaced.yml.html`) - You can also generate a PDF of the slides (by printing them to a file; but be patient with your browser!) .debug[[shared/about-slides.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/about-slides.md)] --- ## These slides are constantly updated - Feel free to check the GitHub repository for updates: https://github.com/jpetazzo/container.training - Look for branches named YYYY-MM-... - You can also find specific decks and other resources on: https://container.training/ .debug[[shared/about-slides.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/about-slides.md)] --- class: extra-details ## Extra details - This slide has a little magnifying glass in the top left corner - This magnifying glass indicates slides that provide extra details - Feel free to skip them if: - you are in a hurry - you are new to this and want to avoid cognitive overload - you want only the most essential information - You can review these slides another time if you want, they'll be waiting for you ☺ .debug[[shared/about-slides.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/about-slides.md)] --- name: toc-part-1 ## Part 1 - [Pre-requirements](#toc-pre-requirements) - [Our sample application](#toc-our-sample-application) - [Identifying bottlenecks](#toc-identifying-bottlenecks) - [SwarmKit](#toc-swarmkit) - [Declarative vs imperative](#toc-declarative-vs-imperative) - [Swarm mode](#toc-swarm-mode) - [Creating our first Swarm](#toc-creating-our-first-swarm) .debug[(auto-generated TOC)] --- name: toc-part-2 ## Part 2 - [Running our first Swarm service](#toc-running-our-first-swarm-service) - [Our app on Swarm](#toc-our-app-on-swarm) - [Hosting our own registry](#toc-hosting-our-own-registry) - [Global scheduling](#toc-global-scheduling) - [Swarm Stacks](#toc-swarm-stacks) - [CI/CD for Docker and orchestration](#toc-cicd-for-docker-and-orchestration) .debug[(auto-generated TOC)] --- name: toc-part-3 ## Part 3 - [Breaking into an overlay network](#toc-breaking-into-an-overlay-network) - [Securing overlay networks](#toc-securing-overlay-networks) - [Updating services](#toc-updating-services) - [Rolling updates](#toc-rolling-updates) - [Health checks and auto-rollbacks](#toc-health-checks-and-auto-rollbacks) - [SwarmKit debugging tools](#toc-swarmkit-debugging-tools) .debug[(auto-generated TOC)] --- name: toc-part-4 ## Part 4 - [Secrets management and encryption at rest](#toc-secrets-management-and-encryption-at-rest) - [Least privilege model](#toc-least-privilege-model) - [Centralized logging](#toc-centralized-logging) - [Setting up ELK to store container logs](#toc-setting-up-elk-to-store-container-logs) - [Metrics collection](#toc-metrics-collection) - [Dealing with stateful services](#toc-dealing-with-stateful-services) - [Controlling Docker from a container](#toc-controlling-docker-from-a-container) - [Node management](#toc-node-management) - [Links and resources](#toc-links-and-resources) .debug[(auto-generated TOC)] .debug[[shared/toc.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/toc.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/Container-Ship-Freighter-Navigation-Elbe-Romance-1782991.jpg)] --- name: toc-pre-requirements class: title Pre-requirements .nav[ [Previous part](#toc-) | [Back to table of contents](#toc-part-1) | [Next part](#toc-our-sample-application) ] .debug[(automatically generated title slide)] --- # Pre-requirements - Be comfortable with the UNIX command line - navigating directories - editing files - a little bit of bash-fu (environment variables, loops) - Some Docker knowledge - `docker run`, `docker ps`, `docker build` - ideally, you know how to write a Dockerfile and build it
(even if it's a `FROM` line and a couple of `RUN` commands) - It's totally OK if you are not a Docker expert! .debug[[shared/prereqs.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/prereqs.md)] --- class: title *Tell me and I forget.*
*Teach me and I remember.*
*Involve me and I learn.* Misattributed to Benjamin Franklin [(Probably inspired by Chinese Confucian philosopher Xunzi)](https://www.barrypopik.com/index.php/new_york_city/entry/tell_me_and_i_forget_teach_me_and_i_may_remember_involve_me_and_i_will_lear/) .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- ## Hands-on sections - There will be *a lot* of examples and demos - We are going to build, ship, and run containers (and sometimes, clusters!) - If you want, you can run all the examples and demos in your environment (but you don't have to; it's up to you!) - All hands-on sections are clearly identified, like the gray rectangle below .lab[ - This is a command that we're gonna run: ```bash echo hello world ``` ] .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- class: in-person ## Where are we going to run our containers? .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- class: in-person, pic ![You get a cluster](images/you-get-a-cluster.jpg) .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- ## If you're attending a live training or workshop - Each person gets a private lab environment (depending on the scenario, this will be one VM, one cluster, multiple clusters...) - The instructor will tell you how to connect to your environment - Your lab environments will be available for the duration of the workshop (check with your instructor to know exactly when they'll be shutdown) .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- ## Running your own lab environments - If you are following a self-paced course... - Or watching a replay of a recorded course... - ...You will need to set up a local environment for the labs - If you want to deliver your own training or workshop: - deployment scripts are available in the [prepare-labs] directory - you can use them to automatically deploy many lab environments - they support many different infrastructure providers [prepare-labs]: https://github.com/jpetazzo/container.training/tree/main/prepare-labs .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- class: in-person ## Why don't we run containers locally? - Installing this stuff can be hard on some machines (32 bits CPU or OS... Laptops without administrator access... etc.) - *"The whole team downloaded all these container images from the WiFi!
... and it went great!"* (Literally no-one ever) - All you need is a computer (or even a phone or tablet!), with: - an Internet connection - a web browser - an SSH client .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- class: in-person ## SSH clients - On Linux, OS X, FreeBSD... you are probably all set - On Windows, get one of these: - [putty](http://www.putty.org/) - Microsoft [Win32 OpenSSH](https://github.com/PowerShell/Win32-OpenSSH/wiki/Install-Win32-OpenSSH) - [Git BASH](https://git-for-windows.github.io/) - [MobaXterm](http://mobaxterm.mobatek.net/) - On Android, [JuiceSSH](https://juicessh.com/) ([Play Store](https://play.google.com/store/apps/details?id=com.sonelli.juicessh)) works pretty well - Nice-to-have: [Mosh](https://mosh.org/) instead of SSH, if your Internet connection tends to lose packets .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- class: in-person, extra-details ## What is this Mosh thing? *You don't have to use Mosh or even know about it to follow along.
We're just telling you about it because some of us think it's cool!* - Mosh is "the mobile shell" - It is essentially SSH over UDP, with roaming features - It retransmits packets quickly, so it works great even on lossy connections (Like hotel or conference WiFi) - It has intelligent local echo, so it works great even in high-latency connections (Like hotel or conference WiFi) - It supports transparent roaming when your client IP address changes (Like when you hop from hotel to conference WiFi) .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- class: in-person, extra-details ## Using Mosh - To install it: `(apt|yum|brew) install mosh` - It has been pre-installed on the VMs that we are using - To connect to a remote machine: `mosh user@host` (It is going to establish an SSH connection, then hand off to UDP) - It requires UDP ports to be open (By default, it uses a UDP port between 60000 and 61000) .debug[[shared/handson.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/handson.md)] --- class: in-person ## Testing the connection to our lab environment .lab[ - Connect to your lab environment with your SSH client: ```bash ssh `user`@`A.B.C.D` ssh -p `32323` `user`@`A.B.C.D` ``` (Make sure to replace the highlighted values with the ones provided to you!) ] You should see a prompt looking like this: ``` [A.B.C.D] (...) user@machine ~ $ ``` If anything goes wrong — ask for help! .debug[[shared/connecting.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/connecting.md)] --- class: in-person ## `tailhist` - The shell history of the instructor is available online in real time - The instructor will provide you a "magic URL" (typically, the instructor's lab address on port 1088 or 30088) - Open that URL in your browser and you should see the history - The history is updated in real time (using a WebSocket connection) - It should be green when the WebSocket is connected (if it turns red, reloading the page should fix it) .debug[[shared/connecting.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/connecting.md)] --- ## Doing or re-doing the workshop on your own? - Use something like [Play-With-Docker](http://play-with-docker.com/) or [Play-With-Kubernetes](https://training.play-with-kubernetes.com/) Zero setup effort; but environment are short-lived and might have limited resources - Create your own cluster (local or cloud VMs) Small setup effort; small cost; flexible environments - Create a bunch of clusters for you and your friends ([instructions](https://github.com/jpetazzo/container.training/tree/main/prepare-labs)) Bigger setup effort; ideal for group training .debug[[shared/connecting.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/connecting.md)] --- ## For a consistent Kubernetes experience ... - If you are using your own Kubernetes cluster, you can use [jpetazzo/shpod](https://github.com/jpetazzo/shpod) - `shpod` provides a shell running in a pod on your own cluster - It comes with many tools pre-installed (helm, stern...) - These tools are used in many demos and exercises in these slides - `shpod` also gives you completion and a fancy prompt - It can also be used as an SSH server if needed .debug[[shared/connecting.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/connecting.md)] --- class: self-paced ## Get your own Docker nodes - If you already have some Docker nodes: great! - If not: let's get some thanks to Play-With-Docker .lab[ - Go to http://www.play-with-docker.com/ - Log in - Create your first node ] You will need a Docker ID to use Play-With-Docker. (Creating a Docker ID is free.) .debug[[shared/connecting.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/connecting.md)] --- ## We don't need to connect to ALL the nodes - If your cluster has multiple nodes (e.g. `node1`, `node2`, ...): unless instructed, **all commands must be run from the first node** - We don't need to check out/copy code or manifests on other nodes - During normal operations, we do not need access to the other nodes (but we could log into these nodes to troubleshoot or examine stuff) .debug[[shared/connecting.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/connecting.md)] --- ## Terminals Once in a while, the instructions will say:
"Open a new terminal." There are multiple ways to do this: - create a new window or tab on your machine, and SSH into the VM; - use screen or tmux on the VM and open a new window from there. You are welcome to use the method that you feel the most comfortable with. .debug[[shared/connecting.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/connecting.md)] --- ## Tmux cheat sheet (basic) [Tmux](https://en.wikipedia.org/wiki/Tmux) is a terminal multiplexer like `screen`. *You don't have to use it or even know about it to follow along.
But some of us like to use it to switch between terminals.
It has been preinstalled on your workshop nodes.* - You can start a new session with `tmux`
(or resume or share an existing session with `tmux attach`) - Then use these keyboard shortcuts: - Ctrl-b c → creates a new window - Ctrl-b n → go to next window - Ctrl-b p → go to previous window - Ctrl-b " → split window top/bottom - Ctrl-b % → split window left/right - Ctrl-b arrows → navigate within split windows .debug[[shared/connecting.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/connecting.md)] --- ## Tmux cheat sheet (advanced) - Ctrl-b d → detach session
(resume it later with `tmux attach`) - Ctrl-b Alt-1 → rearrange windows in columns - Ctrl-b Alt-2 → rearrange windows in rows - Ctrl-b , → rename window - Ctrl-b Ctrl-o → cycle pane position (e.g. switch top/bottom) - Ctrl-b PageUp → enter scrollback mode
(use PageUp/PageDown to scroll; Ctrl-c or Enter to exit scrollback) .debug[[shared/connecting.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/connecting.md)] --- ## Brand new versions! - Engine 18.09 - Compose 1.23 - Machine 0.16 .lab[ - Check all installed versions: ```bash docker version docker-compose -v docker-machine -v ``` ] .debug[[swarm/versions.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/versions.md)] --- ## Wait, what, 18.09 ?!? -- - Docker 1.13 = Docker 17.03 (year.month, like Ubuntu) - Every month, there is a new "edge" release (with new features) - Every quarter, there is a new "stable" release - Docker CE releases are maintained 4+ months - Docker EE releases are maintained 12+ months - For more details, check the [Docker EE announcement blog post](https://blog.docker.com/2017/03/docker-enterprise-edition/) .debug[[swarm/versions.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/versions.md)] --- class: extra-details ## Docker CE vs Docker EE - Docker EE: - $$$ - certification for select distros, clouds, and plugins - advanced management features (fine-grained access control, security scanning...) - Docker CE: - free - available through Docker Mac, Docker Windows, and major Linux distros - perfect for individuals and small organizations .debug[[swarm/versions.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/versions.md)] --- class: extra-details ## Why? - More readable for enterprise users (i.e. the very nice folks who are kind enough to pay us big $$$ for our stuff) - No impact for the community (beyond CE/EE suffix and version numbering change) - Both trains leverage the same open source components (containerd, libcontainer, SwarmKit...) - More predictable release schedule (see next slide) .debug[[swarm/versions.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/versions.md)] --- class: pic ![Docker CE/EE release cycle](images/docker-ce-ee-lifecycle.png) .debug[[swarm/versions.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/versions.md)] --- ## What was added when? |||| | ---- | ----- | --- | | 2015 | 1.9 | Overlay (multi-host) networking, network/IPAM plugins | 2016 | 1.10 | Embedded dynamic DNS | 2016 | 1.11 | DNS round robin load balancing | 2016 | 1.12 | Swarm mode, routing mesh, encrypted networking, healthchecks | 2017 | 1.13 | Stacks, attachable overlays, image squash and compress | 2017 | 1.13 | Windows Server 2016 Swarm mode | 2017 | 17.03 | Secrets, encrypted Raft | 2017 | 17.04 | Update rollback, placement preferences (soft constraints) | 2017 | 17.06 | Swarm configs, node/service events, multi-stage build, service logs | 2017 | 17.06 | Windows Server 2016 Swarm overlay networks, secrets | 2017 | 17.09 | ADD/COPY chown, start\_period, stop-signal, overlay2 default | 2017 | 17.12 | containerd, Hyper-V isolation, Windows routing mesh | 2018 | 18.03 | Templates for secrets/configs, multi-yaml stacks, LCOW | 2018 | 18.03 | Stack deploy to Kubernetes, `docker trust`, tmpfs, manifest CLI .debug[[swarm/versions.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/versions.md)] --- name: part-1 class: title, self-paced Part 1 .debug[[swarm-selfpaced.yml](https://github.com/jpetazzo/container.training/tree/main/slides/swarm-selfpaced.yml)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/ShippingContainerSFBay.jpg)] --- name: toc-our-sample-application class: title Our sample application .nav[ [Previous part](#toc-pre-requirements) | [Back to table of contents](#toc-part-1) | [Next part](#toc-identifying-bottlenecks) ] .debug[(automatically generated title slide)] --- # Our sample application - We will clone the GitHub repository onto our `node1` - The repository also contains scripts and tools that we will use through the workshop .lab[ - Clone the repository on `node1`: ```bash git clone https://github.com/jpetazzo/container.training ``` ] (You can also fork the repository on GitHub and clone your fork if you prefer that.) .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## Downloading and running the application Let's start this before we look around, as downloading will take a little time... .lab[ - Go to the `dockercoins` directory, in the cloned repository: ```bash cd ~/container.training/dockercoins ``` - Use Compose to build and run all containers: ```bash docker-compose up ``` ] Compose tells Docker to build all container images (pulling the corresponding base images), then starts all containers, and displays aggregated logs. .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## What's this application? -- - It is a DockerCoin miner! 💰🐳📦🚢 -- - No, you can't buy coffee with DockerCoin -- - How dockercoins works: - generate a few random bytes - hash these bytes - increment a counter (to keep track of speed) - repeat forever! -- - DockerCoin is *not* a cryptocurrency (the only common points are "randomness," "hashing," and "coins" in the name) .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## DockerCoin in the microservices era - The dockercoins app is made of 5 services: - `rng` = web service generating random bytes - `hasher` = web service computing hash of POSTed data - `worker` = background process calling `rng` and `hasher` - `webui` = web interface to watch progress - `redis` = data store (holds a counter updated by `worker`) - These 5 services are visible in the application's Compose file, [docker-compose.yml]( https://github.com/jpetazzo/container.training/blob/master/dockercoins/docker-compose.yml) .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## How dockercoins works - `worker` invokes web service `rng` to generate random bytes - `worker` invokes web service `hasher` to hash these bytes - `worker` does this in an infinite loop - every second, `worker` updates `redis` to indicate how many loops were done - `webui` queries `redis`, and computes and exposes "hashing speed" in our browser *(See diagram on next slide!)* .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- class: pic ![Diagram showing the 5 containers of the applications](images/dockercoins-diagram.png) .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## Service discovery in container-land How does each service find out the address of the other ones? -- - We do not hard-code IP addresses in the code - We do not hard-code FQDNs in the code, either - We just connect to a service name, and container-magic does the rest (And by container-magic, we mean "a crafty, dynamic, embedded DNS server") .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## Example in `worker/worker.py` ```python redis = Redis("`redis`") def get_random_bytes(): r = requests.get("http://`rng`/32") return r.content def hash_bytes(data): r = requests.post("http://`hasher`/", data=data, headers={"Content-Type": "application/octet-stream"}) ``` (Full source code available [here]( https://github.com/jpetazzo/container.training/blob/8279a3bce9398f7c1a53bdd95187c53eda4e6435/dockercoins/worker/worker.py#L17 )) .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- class: extra-details ## Links, naming, and service discovery - Containers can have network aliases (resolvable through DNS) - Compose file version 2+ makes each container reachable through its service name - Compose file version 1 required "links" sections to accomplish this - Network aliases are automatically namespaced - you can have multiple apps declaring and using a service named `database` - containers in the blue app will resolve `database` to the IP of the blue database - containers in the green app will resolve `database` to the IP of the green database .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## Show me the code! - You can check the GitHub repository with all the materials of this workshop:
https://github.com/jpetazzo/container.training - The application is in the [dockercoins]( https://github.com/jpetazzo/container.training/tree/master/dockercoins) subdirectory - The Compose file ([docker-compose.yml]( https://github.com/jpetazzo/container.training/blob/master/dockercoins/docker-compose.yml)) lists all 5 services - `redis` is using an official image from the Docker Hub - `hasher`, `rng`, `worker`, `webui` are each built from a Dockerfile - Each service's Dockerfile and source code is in its own directory (`hasher` is in the [hasher](https://github.com/jpetazzo/container.training/blob/master/dockercoins/hasher/) directory, `rng` is in the [rng](https://github.com/jpetazzo/container.training/blob/master/dockercoins/rng/) directory, etc.) .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- class: extra-details ## Compose file format version *This is relevant only if you have used Compose before 2016...* - Compose 1.6 introduced support for a new Compose file format (aka "v2") - Services are no longer at the top level, but under a `services` section - There has to be a `version` key at the top level, with value `"2"` (as a string, not an integer) - Containers are placed on a dedicated network, making links unnecessary - There are other minor differences, but upgrade is easy and straightforward .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## Our application at work - On the left-hand side, the "rainbow strip" shows the container names - On the right-hand side, we see the output of our containers - We can see the `worker` service making requests to `rng` and `hasher` - For `rng` and `hasher`, we see HTTP access logs .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## Connecting to the web UI - "Logs are exciting and fun!" (No-one, ever) - The `webui` container exposes a web dashboard; let's view it .lab[ - With a web browser, connect to `node1` on port 8000 - Remember: the `nodeX` aliases are valid only on the nodes themselves - In your browser, you need to enter the IP address of your node ] A drawing area should show up, and after a few seconds, a blue graph will appear. .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- class: self-paced, extra-details ## If the graph doesn't load If you just see a `Page not found` error, it might be because your Docker Engine is running on a different machine. This can be the case if: - you are using the Docker Toolbox - you are using a VM (local or remote) created with Docker Machine - you are controlling a remote Docker Engine When you run DockerCoins in development mode, the web UI static files are mapped to the container using a volume. Alas, volumes can only work on a local environment, or when using Docker Desktop for Mac or Windows. How to fix this? Stop the app with `^C`, edit `dockercoins.yml`, comment out the `volumes` section, and try again. .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- class: extra-details ## Why does the speed seem irregular? - It *looks like* the speed is approximately 4 hashes/second - Or more precisely: 4 hashes/second, with regular dips down to zero - Why? -- class: extra-details - The app actually has a constant, steady speed: 3.33 hashes/second
(which corresponds to 1 hash every 0.3 seconds, for *reasons*) - Yes, and? .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- class: extra-details ## The reason why this graph is *not awesome* - The worker doesn't update the counter after every loop, but up to once per second - The speed is computed by the browser, checking the counter about once per second - Between two consecutive updates, the counter will increase either by 4, or by 0 - The perceived speed will therefore be 4 - 4 - 4 - 0 - 4 - 4 - 0 etc. - What can we conclude from this? -- class: extra-details - "I'm clearly incapable of writing good frontend code!" 😀 — Jérôme .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## Stopping the application - If we interrupt Compose (with `^C`), it will politely ask the Docker Engine to stop the app - The Docker Engine will send a `TERM` signal to the containers - If the containers do not exit in a timely manner, the Engine sends a `KILL` signal .lab[ - Stop the application by hitting `^C` ] -- Some containers exit immediately, others take longer. The containers that do not handle `SIGTERM` end up being killed after a 10s timeout. If we are very impatient, we can hit `^C` a second time! .debug[[shared/sampleapp.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/sampleapp.md)] --- ## Restarting in the background - Many flags and commands of Compose are modeled after those of `docker` .lab[ - Start the app in the background with the `-d` option: ```bash docker-compose up -d ``` - Check that our app is running with the `ps` command: ```bash docker-compose ps ``` ] `docker-compose ps` also shows the ports exposed by the application. .debug[[shared/composescale.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composescale.md)] --- class: extra-details ## Viewing logs - The `docker-compose logs` command works like `docker logs` .lab[ - View all logs since container creation and exit when done: ```bash docker-compose logs ``` - Stream container logs, starting at the last 10 lines for each container: ```bash docker-compose logs --tail 10 --follow ``` ] Tip: use `^S` and `^Q` to pause/resume log output. .debug[[shared/composescale.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composescale.md)] --- ## Scaling up the application - Our goal is to make that performance graph go up (without changing a line of code!) -- - Before trying to scale the application, we'll figure out if we need more resources (CPU, RAM...) - For that, we will use good old UNIX tools on our Docker node .debug[[shared/composescale.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composescale.md)] --- ## Looking at resource usage - Let's look at CPU, memory, and I/O usage .lab[ - run `top` to see CPU and memory usage (you should see idle cycles) - run `vmstat 1` to see I/O usage (si/so/bi/bo)
(the 4 numbers should be almost zero, except `bo` for logging) ] We have available resources. - Why? - How can we use them? .debug[[shared/composescale.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composescale.md)] --- ## Scaling workers on a single node - Docker Compose supports scaling - Let's scale `worker` and see what happens! .lab[ - Start one more `worker` container: ```bash docker-compose up -d --scale worker=2 ``` - Look at the performance graph (it should show a x2 improvement) - Look at the aggregated logs of our containers (`worker_2` should show up) - Look at the impact on CPU load with e.g. top (it should be negligible) ] .debug[[shared/composescale.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composescale.md)] --- ## Adding more workers - Great, let's add more workers and call it a day, then! .lab[ - Start eight more `worker` containers: ```bash docker-compose up -d --scale worker=10 ``` - Look at the performance graph: does it show a x10 improvement? - Look at the aggregated logs of our containers - Look at the impact on CPU load and memory usage ] .debug[[shared/composescale.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composescale.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/aerial-view-of-containers.jpg)] --- name: toc-identifying-bottlenecks class: title Identifying bottlenecks .nav[ [Previous part](#toc-our-sample-application) | [Back to table of contents](#toc-part-1) | [Next part](#toc-swarmkit) ] .debug[(automatically generated title slide)] --- # Identifying bottlenecks - You should have seen a 3x speed bump (not 10x) - Adding workers didn't result in linear improvement - *Something else* is slowing us down -- - ... But what? -- - The code doesn't have instrumentation - Let's use state-of-the-art HTTP performance analysis!
(i.e. good old tools like `ab`, `httping`...) .debug[[shared/composescale.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composescale.md)] --- ## Accessing internal services - `rng` and `hasher` are exposed on ports 8001 and 8002 - This is declared in the Compose file: ```yaml ... rng: build: rng ports: - "8001:80" hasher: build: hasher ports: - "8002:80" ... ``` .debug[[shared/composescale.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composescale.md)] --- ## Measuring latency under load We will use `httping`. .lab[ - Check the latency of `rng`: ```bash httping -c 3 localhost:8001 ``` - Check the latency of `hasher`: ```bash httping -c 3 localhost:8002 ``` ] `rng` has a much higher latency than `hasher`. .debug[[shared/composescale.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composescale.md)] --- ## Let's draw hasty conclusions - The bottleneck seems to be `rng` - *What if* we don't have enough entropy and can't generate enough random numbers? - We need to scale out the `rng` service on multiple machines! Note: this is a fiction! We have enough entropy. But we need a pretext to scale out. (In fact, the code of `rng` uses `/dev/urandom`, which never runs out of entropy...
...and is [just as good as `/dev/random`](http://www.slideshare.net/PacSecJP/filippo-plain-simple-reality-of-entropy).) .debug[[shared/hastyconclusions.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/hastyconclusions.md)] --- ## Clean up - Before moving on, let's remove those containers .lab[ - Tell Compose to remove everything: ```bash docker-compose down ``` ] .debug[[shared/composedown.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/composedown.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/blue-containers.jpg)] --- name: toc-swarmkit class: title SwarmKit .nav[ [Previous part](#toc-identifying-bottlenecks) | [Back to table of contents](#toc-part-1) | [Next part](#toc-declarative-vs-imperative) ] .debug[(automatically generated title slide)] --- # SwarmKit - [SwarmKit](https://github.com/docker/swarmkit) is an open source toolkit to build multi-node systems - It is a reusable library, like libcontainer, libnetwork, vpnkit ... - It is a plumbing part of the Docker ecosystem -- .footnote[🐳 Did you know that кит means "whale" in Russian?] .debug[[swarm/swarmkit.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmkit.md)] --- ## SwarmKit features - Highly-available, distributed store based on [Raft]( https://en.wikipedia.org/wiki/Raft_%28computer_science%29)
(avoids depending on an external store: easier to deploy; higher performance) - Dynamic reconfiguration of Raft without interrupting cluster operations - *Services* managed with a *declarative API*
(implementing *desired state* and *reconciliation loop*) - Integration with overlay networks and load balancing - Strong emphasis on security: - automatic TLS keying and signing; automatic cert rotation - full encryption of the data plane; automatic key rotation - least privilege architecture (single-node compromise ≠ cluster compromise) - on-disk encryption with optional passphrase .debug[[swarm/swarmkit.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmkit.md)] --- class: extra-details ## Where is the key/value store? - Many orchestration systems use a key/value store backed by a consensus algorithm
(k8s→etcd→Raft, mesos→zookeeper→ZAB, etc.) - SwarmKit implements the Raft algorithm directly
(Nomad is similar; thanks [@cbednarski](https://twitter.com/@cbednarski), [@diptanu](https://twitter.com/diptanu) and others for pointing it out!) - Analogy courtesy of [@aluzzardi](https://twitter.com/aluzzardi): *It's like B-Trees and RDBMS. They are different layers, often associated. But you don't need to bring up a full SQL server when all you need is to index some data.* - As a result, the orchestrator has direct access to the data
(the main copy of the data is stored in the orchestrator's memory) - Simpler, easier to deploy and operate; also faster .debug[[swarm/swarmkit.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmkit.md)] --- ## SwarmKit concepts (1/2) - A *cluster* will be at least one *node* (preferably more) - A *node* can be a *manager* or a *worker* - A *manager* actively takes part in the Raft consensus, and keeps the Raft log - You can talk to a *manager* using the SwarmKit API - One *manager* is elected as the *leader*; other managers merely forward requests to it - The *workers* get their instructions from the *managers* - Both *workers* and *managers* can run containers .debug[[swarm/swarmkit.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmkit.md)] --- ## Illustration On the next slide: - whales = nodes (workers and managers) - monkeys = managers - purple monkey = leader - grey monkeys = followers - dotted triangle = raft protocol .debug[[swarm/swarmkit.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmkit.md)] --- class: pic ![Illustration](images/swarm-mode.svg) .debug[[swarm/swarmkit.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmkit.md)] --- ## SwarmKit concepts (2/2) - The *managers* expose the SwarmKit API - Using the API, you can indicate that you want to run a *service* - A *service* is specified by its *desired state*: which image, how many instances... - The *leader* uses different subsystems to break down services into *tasks*:
orchestrator, scheduler, allocator, dispatcher - A *task* corresponds to a specific container, assigned to a specific *node* - *Nodes* know which *tasks* should be running, and will start or stop containers accordingly (through the Docker Engine API) You can refer to the [NOMENCLATURE](https://github.com/docker/swarmkit/blob/master/design/nomenclature.md) in the SwarmKit repo for more details. .debug[[swarm/swarmkit.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmkit.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/chinook-helicopter-container.jpg)] --- name: toc-declarative-vs-imperative class: title Declarative vs imperative .nav[ [Previous part](#toc-swarmkit) | [Back to table of contents](#toc-part-1) | [Next part](#toc-swarm-mode) ] .debug[(automatically generated title slide)] --- # Declarative vs imperative - Our container orchestrator puts a very strong emphasis on being *declarative* - Declarative: *I would like a cup of tea.* - Imperative: *Boil some water. Pour it in a teapot. Add tea leaves. Steep for a while. Serve in a cup.* -- - Declarative seems simpler at first ... -- - ... As long as you know how to brew tea .debug[[shared/declarative.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/declarative.md)] --- ## Declarative vs imperative - What declarative would really be: *I want a cup of tea, obtained by pouring an infusion¹ of tea leaves in a cup.* -- *¹An infusion is obtained by letting the object steep a few minutes in hot² water.* -- *²Hot liquid is obtained by pouring it in an appropriate container³ and setting it on a stove.* -- *³Ah, finally, containers! Something we know about. Let's get to work, shall we?* -- .footnote[Did you know there was an [ISO standard](https://en.wikipedia.org/wiki/ISO_3103) specifying how to brew tea?] .debug[[shared/declarative.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/declarative.md)] --- ## Declarative vs imperative - Imperative systems: - simpler - if a task is interrupted, we have to restart from scratch - Declarative systems: - if a task is interrupted (or if we show up to the party half-way through), we can figure out what's missing and do only what's necessary - we need to be able to *observe* the system - ... and compute a "diff" between *what we have* and *what we want* .debug[[shared/declarative.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/declarative.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/container-cranes.jpg)] --- name: toc-swarm-mode class: title Swarm mode .nav[ [Previous part](#toc-declarative-vs-imperative) | [Back to table of contents](#toc-part-1) | [Next part](#toc-creating-our-first-swarm) ] .debug[(automatically generated title slide)] --- # Swarm mode - Since version 1.12, the Docker Engine embeds SwarmKit - All the SwarmKit features are "asleep" until you enable "Swarm mode" - Examples of Swarm Mode commands: - `docker swarm` (enable Swarm mode; join a Swarm; adjust cluster parameters) - `docker node` (view nodes; promote/demote managers; manage nodes) - `docker service` (create and manage services) ??? - The Docker API exposes the same concepts - The SwarmKit API is also exposed (on a separate socket) .debug[[swarm/swarmmode.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmmode.md)] --- ## Swarm mode needs to be explicitly activated - By default, all this new code is inactive - Swarm mode can be enabled, "unlocking" SwarmKit functions
(services, out-of-the-box overlay networks, etc.) .lab[ - Try a Swarm-specific command: ```bash docker node ls ``` ] -- You will get an error message: ``` Error response from daemon: This node is not a swarm manager. [...] ``` .debug[[swarm/swarmmode.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmmode.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/container-housing.jpg)] --- name: toc-creating-our-first-swarm class: title Creating our first Swarm .nav[ [Previous part](#toc-swarm-mode) | [Back to table of contents](#toc-part-1) | [Next part](#toc-running-our-first-swarm-service) ] .debug[(automatically generated title slide)] --- # Creating our first Swarm - The cluster is initialized with `docker swarm init` - This should be executed on a first, seed node - .warning[DO NOT execute `docker swarm init` on multiple nodes!] You would have multiple disjoint clusters. .lab[ - Create our cluster from node1: ```bash docker swarm init ``` ] -- class: advertise-addr If Docker tells you that it `could not choose an IP address to advertise`, see next slide! .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: advertise-addr ## IP address to advertise - When running in Swarm mode, each node *advertises* its address to the others
(i.e. it tells them *"you can contact me on 10.1.2.3:2377"*) - If the node has only one IP address, it is used automatically
(The addresses of the loopback interface and the Docker bridge are ignored) - If the node has multiple IP addresses, you **must** specify which one to use
(Docker refuses to pick one randomly) - You can specify an IP address or an interface name
(in the latter case, Docker will read the IP address of the interface and use it) - You can also specify a port number
(otherwise, the default port 2377 will be used) .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: advertise-addr ## Using a non-default port number - Changing the *advertised* port does not change the *listening* port - If you only pass `--advertise-addr eth0:7777`, Swarm will still listen on port 2377 - You will probably need to pass `--listen-addr eth0:7777` as well - This is to accommodate scenarios where these ports *must* be different
(port mapping, load balancers...) Example to run Swarm on a different port: ```bash docker swarm init --advertise-addr eth0:7777 --listen-addr eth0:7777 ``` .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: advertise-addr ## Which IP address should be advertised? - If your nodes have only one IP address, it's safe to let autodetection do the job .small[(Except if your instances have different private and public addresses, e.g. on EC2, and you are building a Swarm involving nodes inside and outside the private network: then you should advertise the public address.)] - If your nodes have multiple IP addresses, pick an address which is reachable *by every other node* of the Swarm - If you are using [play-with-docker](http://play-with-docker.com/), use the IP address shown next to the node name .small[(This is the address of your node on your private internal overlay network. The other address that you might see is the address of your node on the `docker_gwbridge` network, which is used for outbound traffic.)] Examples: ```bash docker swarm init --advertise-addr 172.24.0.2 docker swarm init --advertise-addr eth0 ``` .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: extra-details ## Using a separate interface for the data path - You can use different interfaces (or IP addresses) for control and data - You set the _control plane path_ with `--advertise-addr` and `--listen-addr` (This will be used for SwarmKit manager/worker communication, leader election, etc.) - You set the _data plane path_ with `--data-path-addr` (This will be used for traffic between containers) - Both flags can accept either an IP address, or an interface name (When specifying an interface name, Docker will use its first IP address) .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- ## Token generation - In the output of `docker swarm init`, we have a message confirming that our node is now the (single) manager: ``` Swarm initialized: current node (8jud...) is now a manager. ``` - Docker generated two security tokens (like passphrases or passwords) for our cluster - The CLI shows us the command to use on other nodes to add them to the cluster using the "worker" security token: ``` To add a worker to this swarm, run the following command: docker swarm join \ --token SWMTKN-1-59fl4ak4nqjmao1ofttrc4eprhrola2l87... \ 172.31.4.182:2377 ``` .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: extra-details ## Checking that Swarm mode is enabled .lab[ - Run the traditional `docker info` command: ```bash docker info ``` ] The output should include: ``` Swarm: active NodeID: 8jud7o8dax3zxbags3f8yox4b Is Manager: true ClusterID: 2vcw2oa9rjps3a24m91xhvv0c ... ``` .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- ## Running our first Swarm mode command - Let's retry the exact same command as earlier .lab[ - List the nodes (well, the only node) of our cluster: ```bash docker node ls ``` ] The output should look like the following: ``` ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS 8jud...ox4b * node1 Ready Active Leader ``` .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- ## Adding nodes to the Swarm - A cluster with one node is not a lot of fun - Let's add `node2`! - We need the token that was shown earlier -- - You wrote it down, right? -- - Don't panic, we can easily see it again 😏 .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- ## Adding nodes to the Swarm .lab[ - Show the token again: ```bash docker swarm join-token worker ``` - Log into `node2`: ```bash ssh node2 ``` - Copy-paste the `docker swarm join ...` command
(that was displayed just before) ] .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: extra-details ## Check that the node was added correctly - Stay on `node2` for now! .lab[ - We can still use `docker info` to verify that the node is part of the Swarm: ```bash docker info | grep ^Swarm ``` ] - However, Swarm commands will not work; try, for instance: ```bash docker node ls ``` - This is because the node that we added is currently a *worker* - Only *managers* can accept Swarm-specific commands .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- ## View our two-node cluster - Let's go back to `node1` and see what our cluster looks like .lab[ - Switch back to `node1` (with `exit`, `Ctrl-D` ...) - View the cluster from `node1`, which is a manager: ```bash docker node ls ``` ] The output should be similar to the following: ``` ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS 8jud...ox4b * node1 Ready Active Leader ehb0...4fvx node2 Ready Active ``` .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: under-the-hood ## Under the hood: docker swarm init When we do `docker swarm init`: - a keypair is created for the root CA of our Swarm - a keypair is created for the first node - a certificate is issued for this node - the join tokens are created .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: under-the-hood ## Under the hood: join tokens There is one token to *join as a worker*, and another to *join as a manager*. The join tokens have two parts: - a secret key (preventing unauthorized nodes from joining) - a fingerprint of the root CA certificate (preventing MITM attacks) If a token is compromised, it can be rotated instantly with: ``` docker swarm join-token --rotate
``` .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: under-the-hood ## Under the hood: docker swarm join When a node joins the Swarm: - it is issued its own keypair, signed by the root CA - if the node is a manager: - it joins the Raft consensus - it connects to the current leader - it accepts connections from worker nodes - if the node is a worker: - it connects to one of the managers (leader or follower) .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: under-the-hood ## Under the hood: cluster communication - The *control plane* is encrypted with AES-GCM; keys are rotated every 12 hours - Authentication is done with mutual TLS; certificates are rotated every 90 days (`docker swarm update` allows to change this delay or to use an external CA) - The *data plane* (communication between containers) is not encrypted by default (but this can be activated on a by-network basis, using IPSEC, leveraging hardware crypto if available) .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- class: under-the-hood ## Under the hood: I want to know more! Revisit SwarmKit concepts: - Docker 1.12 Swarm Mode Deep Dive Part 1: Topology ([video](https://www.youtube.com/watch?v=dooPhkXT9yI)) - Docker 1.12 Swarm Mode Deep Dive Part 2: Orchestration ([video](https://www.youtube.com/watch?v=_F6PSP-qhdA)) Some presentations from the Docker Distributed Systems Summit in Berlin: - Heart of the SwarmKit: Topology Management ([slides](https://speakerdeck.com/aluzzardi/heart-of-the-swarmkit-topology-management)) - Heart of the SwarmKit: Store, Topology & Object Model ([slides](https://www.slideshare.net/Docker/heart-of-the-swarmkit-store-topology-object-model)) ([video](https://www.youtube.com/watch?v=EmePhjGnCXY)) And DockerCon Black Belt talks: .blackbelt[DC17US: Everything You Thought You Already Knew About Orchestration ([video](https://www.youtube.com/watch?v=Qsv-q8WbIZY&list=PLkA60AVN3hh-biQ6SCtBJ-WVTyBmmYho8&index=6))] .blackbelt[DC17EU: Container Orchestration from Theory to Practice ([video](https://dockercon.docker.com/watch/5fhwnQxW8on1TKxPwwXZ5r))] .debug[[swarm/creatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/creatingswarm.md)] --- ## Adding more manager nodes - Right now, we have only one manager (node1) - If we lose it, we lose quorum - and that's *very bad!* - Containers running on other nodes will be fine ... - But we won't be able to get or set anything related to the cluster - If the manager is permanently gone, we will have to do a manual repair! - Nobody wants to do that ... so let's make our cluster highly available .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- class: self-paced ## Adding more managers With Play-With-Docker: ```bash TOKEN=$(docker swarm join-token -q manager) for N in $(seq 3 5); do export DOCKER_HOST=tcp://node$N:2375 docker swarm join --token $TOKEN node1:2377 done unset DOCKER_HOST ``` .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- class: in-person ## Building our full cluster - Let's get the token, and use a one-liner for the remaining node with SSH .lab[ - Obtain the manager token: ```bash TOKEN=$(docker swarm join-token -q manager) ``` - Add the remaining node: ```bash ssh node3 docker swarm join --token $TOKEN node1:2377 ``` ] [That was easy.](https://www.youtube.com/watch?v=3YmMNpbFjp0) .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- ## Controlling the Swarm from other nodes .lab[ - Try the following command on a few different nodes: ```bash docker node ls ``` ] On manager nodes:
you will see the list of nodes, with a `*` denoting the node you're talking to. On non-manager nodes:
you will get an error message telling you that the node is not a manager. As we saw earlier, you can only control the Swarm through a manager node. .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- class: self-paced ## Play-With-Docker node status icon - If you're using Play-With-Docker, you get node status icons - Node status icons are displayed left of the node name - No icon = no Swarm mode detected - Solid blue icon = Swarm manager detected - Blue outline icon = Swarm worker detected ![Play-With-Docker icons](images/pwd-icons.png) .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- ## Dynamically changing the role of a node - We can change the role of a node on the fly: `docker node promote nodeX` → make nodeX a manager
`docker node demote nodeX` → make nodeX a worker .lab[ - See the current list of nodes: ``` docker node ls ``` - Promote any worker node to be a manager: ``` docker node promote
``` ] .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- ## How many managers do we need? - 2N+1 nodes can (and will) tolerate N failures
(you can have an even number of managers, but there is no point) -- - 1 manager = no failure - 3 managers = 1 failure - 5 managers = 2 failures (or 1 failure during 1 maintenance) - 7 managers and more = now you might be overdoing it for most designs .footnote[ see [Docker's admin guide](https://docs.docker.com/engine/swarm/admin_guide/#add-manager-nodes-for-fault-tolerance) on node failure and datacenter redundancy ] .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- ## Why not have *all* nodes be managers? - With Raft, writes have to go to (and be acknowledged by) all nodes - Thus, it's harder to reach consensus in larger groups - Only one manager is Leader (writable), so more managers ≠ more capacity - Managers should be < 10ms latency from each other - These design parameters lead us to recommended designs .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- ## What would McGyver do? - Keep managers in one region (multi-zone/datacenter/rack) - Groups of 3 or 5 nodes: all are managers. Beyond 5, separate out managers and workers - Groups of 10-100 nodes: pick 5 "stable" nodes to be managers - Groups of more than 100 nodes: watch your managers' CPU and RAM - 16GB memory or more, 4 CPU's or more, SSD's for Raft I/O - otherwise, break down your nodes in multiple smaller clusters .footnote[ Cloud pro-tip: use separate auto-scaling groups for managers and workers See docker's "[Running Docker at scale](https://success.docker.com/article/running-docker-ee-at-scale)" document ] .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- ## What's the upper limit? - We don't know! - Internal testing at Docker Inc.: 1000-10000 nodes is fine - deployed to a single cloud region - one of the main take-aways was *"you're gonna need a bigger manager"* - Testing by the community: [4700 heterogeneous nodes all over the 'net](https://sematext.com/blog/2016/11/14/docker-swarm-lessons-from-swarm3k/) - it just works, assuming they have the resources - more nodes require manager CPU and networking; more containers require RAM - scheduling of large jobs (70,000 containers) is slow, though ([getting better](https://github.com/moby/moby/pull/37372)!) .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- ## Real-life deployment methods -- Running commands manually over SSH -- (lol jk) -- - Using your favorite configuration management tool - [Docker for AWS](https://docs.docker.com/docker-for-aws/#quickstart) - [Docker for Azure](https://docs.docker.com/docker-for-azure/) .debug[[swarm/morenodes.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/morenodes.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/containers-by-the-water.jpg)] --- name: toc-running-our-first-swarm-service class: title Running our first Swarm service .nav[ [Previous part](#toc-creating-our-first-swarm) | [Back to table of contents](#toc-part-2) | [Next part](#toc-our-app-on-swarm) ] .debug[(automatically generated title slide)] --- # Running our first Swarm service - How do we run services? Simplified version: `docker run` → `docker service create` .lab[ - Create a service featuring an Alpine container pinging Google resolvers: ```bash docker service create --name pingpong alpine ping 8.8.8.8 ``` - Check the result: ```bash docker service ps pingpong ``` ] .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Checking service logs (New in Docker Engine 17.05) - Just like `docker logs` shows the output of a specific local container ... - ... `docker service logs` shows the output of all the containers of a specific service .lab[ - Check the output of our ping command: ```bash docker service logs pingpong ``` ] Flags `--follow` and `--tail` are available, as well as a few others. Note: by default, when a container is destroyed (e.g. when scaling down), its logs are lost. .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- class: extra-details ## Looking up where our container is running - The `docker service ps` command told us where our container was scheduled .lab[ - Look up the `NODE` on which the container is running: ```bash docker service ps pingpong ``` - If you use Play-With-Docker, switch to that node's tab, or set `DOCKER_HOST` - Otherwise, `ssh` into that node or use `$(eval docker-machine env node...)` ] .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- class: extra-details ## Viewing the logs of the container .lab[ - See that the container is running and check its ID: ```bash docker ps ``` - View its logs: ```bash docker logs containerID ``` - Go back to `node1` afterwards ] .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Scale our service - Services can be scaled in a pinch with the `docker service update` command .lab[ - Scale the service to ensure 2 copies per node: ```bash docker service update pingpong --replicas 6 ``` - Check that we have two containers on the current node: ```bash docker ps ``` ] .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Monitoring deployment progress with `--detach` (New in Docker Engine 17.10) - The CLI monitors commands that create/update/delete services - In effect, `--detach=false` is the default - synchronous operation - the CLI will monitor and display the progress of our request - it exits only when the operation is complete - Ctrl-C to detach at anytime - `--detach=true` - asynchronous operation - the CLI just submits our request - it exits as soon as the request is committed into Raft .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## To `--detach` or not to `--detach` - `--detach=false` - great when experimenting, to see what's going on - also great when orchestrating complex deployments
(when you want to wait for a service to be ready before starting another) - `--detach=true` - great for independent operations that can be parallelized - great in headless scripts (where nobody's watching anyway) .warning[`--detach=true` does not complete *faster*. It just *doesn't wait* for completion.] .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- class: extra-details ## `--detach` over time - Docker Engine 17.10 and later: the default is `--detach=false` - From Docker Engine 17.05 to 17.09: the default is `--detach=true` - Prior to Docker 17.05: `--detach` doesn't exist (You can watch progress with e.g. `watch docker service ps
`) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## `--detach` in action .lab[ - Scale the service to ensure 3 copies per node: ```bash docker service update pingpong --replicas 9 --detach=false ``` - And then to 4 copies per node: ```bash docker service update pingpong --replicas 12 --detach=true ``` ] .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Expose a service - Services can be exposed, with two special properties: - the public port is available on *every node of the Swarm*, - requests coming on the public port are load balanced across all instances. - This is achieved with option `-p/--publish`; as an approximation: `docker run -p → docker service create -p` - If you indicate a single port number, it will be mapped on a port starting at 30000
(vs. 32768 for single container mapping) - You can indicate two port numbers to set the public port number
(just like with `docker run -p`) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Expose ElasticSearch on its default port .lab[ - Create an ElasticSearch service (and give it a name while we're at it): ```bash docker service create --name search --publish 9200:9200 --replicas 5 \ elasticsearch`:2` ``` ] Note: don't forget the **:2**! The latest version of the ElasticSearch image won't start without mandatory configuration. .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Tasks lifecycle - During the deployment, you will be able to see multiple states: - assigned (the task has been assigned to a specific node) - preparing (this mostly means "pulling the image") - starting - running - When a task is terminated (stopped, killed...) it cannot be restarted (A replacement task will be created) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- class: extra-details, pic ![diagram showing what happens during docker service create, courtesy of @aluzzardi](images/docker-service-create.svg) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Test our service - We mapped port 9200 on the nodes, to port 9200 in the containers - Let's try to reach that port! .lab[ - Try the following command: ```bash curl localhost:9200 ``` ] (If you get `Connection refused`: congratulations, you are very fast indeed! Just try again.) ElasticSearch serves a little JSON document with some basic information about this instance; including a randomly-generated super-hero name. .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Test the load balancing - If we repeat our `curl` command multiple times, we will see different names .lab[ - Send 10 requests, and see which instances serve them: ```bash for N in $(seq 1 10); do curl -s localhost:9200 | jq .name done ``` ] Note: if you don't have `jq` on your Play-With-Docker instance, just install it: ``` apk add --no-cache jq ``` .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Load balancing results Traffic is handled by our clusters [routing mesh]( https://docs.docker.com/engine/swarm/ingress/). Each request is served by one of the instances, in rotation. Note: if you try to access the service from your browser, you will probably see the same instance name over and over, because your browser (unlike curl) will try to re-use the same connection. .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- class: pic ![routing mesh](images/ingress-routing-mesh.png) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Under the hood of the routing mesh - Load balancing is done by IPVS - IPVS is a high-performance, in-kernel load balancer - It's been around for a long time (merged in the kernel since 2.4) - Each node runs a local load balancer (Allowing connections to be routed directly to the destination, without extra hops) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Managing inbound traffic There are many ways to deal with inbound traffic on a Swarm cluster. - Put all (or a subset) of your nodes in a DNS `A` record (good for web clients) - Assign your nodes (or a subset) to an external load balancer (ELB, etc.) - Use a virtual IP and make sure that it is assigned to an "alive" node - etc. .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- class: pic ![external LB](images/ingress-lb.png) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Managing HTTP traffic - The TCP routing mesh doesn't parse HTTP headers - If you want to place multiple HTTP services on port 80/443, you need something more - You can set up NGINX or HAProxy on port 80/443 to route connections to the correct Service, but they need to be "Swarm aware" to dynamically update configs -- - Docker EE provides its own [Layer 7 routing](https://docs.docker.com/ee/ucp/interlock/) - Service labels like `com.docker.lb.hosts=
` are detected automatically via Docker API and dynamically update the configuration -- - Two common open source options: - [Traefik](https://traefik.io/) - popular, many features, requires running on managers, needs key/value for HA - [Docker Flow Proxy](http://proxy.dockerflow.com/) - uses HAProxy, made for Swarm by Docker Captain [@vfarcic](https://twitter.com/vfarcic) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- class: btw-labels ## You should use labels - Labels are a great way to attach arbitrary information to services - Examples: - HTTP vhost of a web app or web service - backup schedule for a stateful service - owner of a service (for billing, paging...) - correlate Swarm objects together (services, volumes, configs, secrets, etc.) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Pro-tip for ingress traffic management - It is possible to use *local* networks with Swarm services - This means that you can do something like this: ```bash docker service create --network host --mode global traefik ... ``` (This runs the `traefik` load balancer on each node of your cluster, in the `host` network) - This gives you native performance (no iptables, no proxy, no nothing!) - The load balancer will "see" the clients' IP addresses - But: a container cannot simultaneously be in the `host` network and another network (You will have to route traffic to containers using exposed ports or UNIX sockets) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- class: extra-details ## Using local networks (`host`, `macvlan` ...) - It is possible to connect services to local networks - Using the `host` network is fairly straightforward (With the caveats described on the previous slide) - Other network drivers are a bit more complicated (IP allocation may have to be coordinated between nodes) - See for instance [this guide]( https://docs.docker.com/engine/userguide/networking/get-started-macvlan/ ) to get started on `macvlan` - See [this PR](https://github.com/moby/moby/pull/32981) for more information about local network drivers in Swarm mode .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Visualize container placement - Let's leverage the Docker API! .lab[ - Run this simple-yet-beautiful visualization app: ```bash cd ~/container.training/stacks docker-compose -f visualizer.yml up -d ``` ] .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Connect to the visualization webapp - It runs a web server on port 8080 .lab[ - Point your browser to port 8080 of your node1's public ip (If you use Play-With-Docker, click on the (8080) badge) ] - The webapp updates the display automatically (you don't need to reload the page) - It only shows Swarm services (not standalone containers) - It shows when nodes go down - It has some glitches (it's not Carrier-Grade Enterprise-Compliant ISO-9001 software) .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Why This Is More Important Than You Think - The visualizer accesses the Docker API *from within a container* - This is a common pattern: run container management tools *in containers* - Instead of viewing your cluster, this could take care of logging, metrics, autoscaling ... - We can run it within a service, too! We won't do it yet, but the command would look like: ```bash docker service create \ --mount source=/var/run/docker.sock,type=bind,target=/var/run/docker.sock \ --name viz --constraint node.role==manager ... ``` .footnote[ Credits: the visualization code was written by [Francisco Miranda](https://github.com/maroshii). [Mano Marks](https://twitter.com/manomarks) adapted it to Swarm and maintains it. ] .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- ## Terminate our services - Before moving on, we will remove those services - `docker service rm` can accept multiple services names or IDs - `docker service ls` can accept the `-q` flag - A Shell snippet a day keeps the cruft away .lab[ - Remove all services with this one liner: ```bash docker service ls -q | xargs docker service rm ``` ] .debug[[swarm/firstservice.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/firstservice.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/distillery-containers.jpg)] --- name: toc-our-app-on-swarm class: title Our app on Swarm .nav[ [Previous part](#toc-running-our-first-swarm-service) | [Back to table of contents](#toc-part-2) | [Next part](#toc-hosting-our-own-registry) ] .debug[(automatically generated title slide)] --- # Our app on Swarm In this part, we will: - **build** images for our app, - **ship** these images with a registry, - **run** services using these images. .debug[[swarm/ourapponswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ourapponswarm.md)] --- ## Why do we need to ship our images? - When we do `docker-compose up`, images are built for our services - These images are present only on the local node - We need these images to be distributed on the whole Swarm - The easiest way to achieve that is to use a Docker registry - Once our images are on a registry, we can reference them when creating our services .debug[[swarm/ourapponswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ourapponswarm.md)] --- class: extra-details ## Build, ship, and run, for a single service If we had only one service (built from a `Dockerfile` in the current directory), our workflow could look like this: ``` docker build -t jpetazzo/doublerainbow:v0.1 . docker push jpetazzo/doublerainbow:v0.1 docker service create jpetazzo/doublerainbow:v0.1 ``` We just have to adapt this to our application, which has 4 services! .debug[[swarm/ourapponswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ourapponswarm.md)] --- ## The plan - Build on our local node (`node1`) - Tag images so that they are named `localhost:5000/servicename` - Upload them to a registry - Create services using the images .debug[[swarm/ourapponswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ourapponswarm.md)] --- ## Which registry do we want to use? .small[ - **Docker Hub** - hosted by Docker Inc. - requires an account (free, no credit card needed) - images will be public (unless you pay) - located in AWS EC2 us-east-1 - **Docker Trusted Registry** - self-hosted commercial product - requires a subscription (free 30-day trial available) - images can be public or private - located wherever you want - **Docker open source registry** - self-hosted barebones repository hosting - doesn't require anything - doesn't come with anything either - located wherever you want - **Lots of 3rd party cloud or self-hosted options** - AWS/Azure/Google Container Registry - GitLab, Quay, JFrog ] .debug[[swarm/ourapponswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ourapponswarm.md)] --- class: extra-details ## Using Docker Hub *If we wanted to use the Docker Hub...* - We would log into the Docker Hub: ```bash docker login ``` - And in the following slides, we would use our Docker Hub login (e.g. `jpetazzo`) instead of the registry address (i.e. `127.0.0.1:5000`) .debug[[swarm/ourapponswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ourapponswarm.md)] --- class: extra-details ## Using Docker Trusted Registry *If we wanted to use DTR, we would...* - Make sure we have a Docker Hub account - [Activate a Docker EE subscription]( https://hub.docker.com/enterprise/trial/) - Install DTR on our machines - Use `dtraddress:port/user` instead of the registry address *This is out of the scope of this workshop!* .debug[[swarm/ourapponswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ourapponswarm.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/lots-of-containers.jpg)] --- name: toc-hosting-our-own-registry class: title Hosting our own registry .nav[ [Previous part](#toc-our-app-on-swarm) | [Back to table of contents](#toc-part-2) | [Next part](#toc-global-scheduling) ] .debug[(automatically generated title slide)] --- # Hosting our own registry - We need to run a `registry` container - It will store images and layers to the local filesystem
(but you can add a config file to use S3, Swift, etc.) - Docker *requires* TLS when communicating with the registry - unless for registries on `127.0.0.0/8` (i.e. `localhost`) - or with the Engine flag `--insecure-registry` - Our strategy: publish the registry container on port 5000,
so that it's available through `127.0.0.1:5000` on each node .debug[[swarm/hostingregistry.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/hostingregistry.md)] --- ## Deploying the registry - We will create a single-instance service, publishing its port on the whole cluster .lab[ - Create the registry service: ```bash docker service create --name registry --publish 5000:5000 registry ``` - Now try the following command; it should return `{"repositories":[]}`: ```bash curl 127.0.0.1:5000/v2/_catalog ``` ] .debug[[swarm/hostingregistry.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/hostingregistry.md)] --- ## Testing our local registry - We can retag a small image, and push it to the registry .lab[ - Make sure we have the busybox image, and retag it: ```bash docker pull busybox docker tag busybox 127.0.0.1:5000/busybox ``` - Push it: ```bash docker push 127.0.0.1:5000/busybox ``` ] .debug[[swarm/testingregistry.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/testingregistry.md)] --- ## Checking what's on our local registry - The registry API has endpoints to query what's there .lab[ - Ensure that our busybox image is now in the local registry: ```bash curl http://127.0.0.1:5000/v2/_catalog ``` ] The curl command should now output: ```json {"repositories":["busybox"]} ``` .debug[[swarm/testingregistry.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/testingregistry.md)] --- ## Build, tag, and push our container images - Compose has named our images `dockercoins_XXX` for each service - We need to retag them (to `127.0.0.1:5000/XXX:v1`) and push them .lab[ - Set `REGISTRY` and `TAG` environment variables to use our local registry - And run this little for loop: ```bash cd ~/container.training/dockercoins export REGISTRY=127.0.0.1:5000 export TAG=v0.1 for SERVICE in hasher rng webui worker; do docker build -t $REGISTRY/$SERVICE:$TAG ./$SERVICE docker push $REGISTRY/$SERVICE:$TAG done ``` ] .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Overlay networks - SwarmKit integrates with overlay networks - Networks are created with `docker network create` - Make sure to specify that you want an *overlay* network
(otherwise you will get a local *bridge* network by default) .lab[ - Create an overlay network for our application: ```bash docker network create --driver overlay dockercoins ``` ] .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Viewing existing networks - Let's confirm that our network was created .lab[ - List existing networks: ```bash docker network ls ``` ] .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Can you spot the differences? The networks `dockercoins` and `ingress` are different from the other ones. Can you see how? -- - They are using a different kind of ID, reflecting the fact that they are SwarmKit objects instead of "classic" Docker Engine objects. - Their *scope* is `swarm` instead of `local`. - They are using the overlay driver. .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- class: extra-details ## Caveats .warning[In Docker 1.12, you cannot join an overlay network with `docker run --net ...`.] Starting with version 1.13, you can, if the network was created with the `--attachable` flag. *Why is that?* Placing a container on a network requires allocating an IP address for this container. The allocation must be done by a manager node (worker nodes cannot update Raft data). As a result, `docker run --net ...` requires collaboration with manager nodes. It alters the code path for `docker run`, so it is allowed only under strict circumstances. .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Run the application - First, create the `redis` service; that one is using a Docker Hub image .lab[ - Create the `redis` service: ```bash docker service create --network dockercoins --name redis redis ``` ] .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Run the other services - Then, start the other services one by one - We will use the images pushed previously .lab[ - Start the other services: ```bash export REGISTRY=127.0.0.1:5000 export TAG=v0.1 for SERVICE in hasher rng webui worker; do docker service create --network dockercoins --detach=true \ --name $SERVICE $REGISTRY/$SERVICE:$TAG done ``` ] .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Expose our application web UI - We need to connect to the `webui` service, but it is not publishing any port - Let's reconfigure it to publish a port .lab[ - Update `webui` so that we can connect to it from outside: ```bash docker service update webui --publish-add 8000:80 ``` ] Note: to "de-publish" a port, you would have to specify the container port. (i.e. in that case, `--publish-rm 80`) .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## What happens when we modify a service? - Let's find out what happened to our `webui` service .lab[ - Look at the tasks and containers associated to `webui`: ```bash docker service ps webui ``` ] -- The first version of the service (the one that was not exposed) has been shutdown. It has been replaced by the new version, with port 80 accessible from outside. (This will be discussed with more details in the section about stateful services.) .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Connect to the web UI - The web UI is now available on port 8000, *on all the nodes of the cluster* .lab[ - If you're using Play-With-Docker, just click on the `(8000)` badge - Otherwise, point your browser to any node, on port 8000 ] .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Scaling the application - We can change scaling parameters with `docker update` as well - We will do the equivalent of `docker-compose scale` .lab[ - Bring up more workers: ```bash docker service update worker --replicas 10 ``` - Check the result in the web UI ] You should see the performance peaking at 10 hashes/s (like before). .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/plastic-containers.JPG)] --- name: toc-global-scheduling class: title Global scheduling .nav[ [Previous part](#toc-hosting-our-own-registry) | [Back to table of contents](#toc-part-2) | [Next part](#toc-swarm-stacks) ] .debug[(automatically generated title slide)] --- # Global scheduling - We want to utilize as best as we can the entropy generators on our nodes - We want to run exactly one `rng` instance per node - SwarmKit has a special scheduling mode for that, let's use it - We cannot enable/disable global scheduling on an existing service - We have to destroy and re-create the `rng` service .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Scaling the `rng` service .lab[ - Remove the existing `rng` service: ```bash docker service rm rng ``` - Re-create the `rng` service with *global scheduling*: ```bash docker service create --name rng --network dockercoins --mode global \ $REGISTRY/rng:$TAG ``` - Look at the result in the web UI ] .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- class: extra-details ## Why do we have to re-create the service? - State reconciliation is handled by a *controller* - The controller knows how to "converge" a scaled service spec to another - It doesn't know how to "transform" a scaled service into a global one
(or vice versa) - This might change in the future (after all, it was possible in 1.12 RC!) - As of Docker Engine 18.03, other parameters requiring to `rm`/`create` the service are: - service name - hostname .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## Removing everything - Before moving on, let's get a clean slate .lab[ - Remove *all* the services: ```bash docker service ls -q | xargs docker service rm ``` ] .debug[[swarm/btp-manual.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/btp-manual.md)] --- ## How did we make our app "Swarm-ready"? This app was written in June 2015. (One year before Swarm mode was released.) What did we change to make it compatible with Swarm mode? -- .lab[ - Go to the app directory: ```bash cd ~/container.training/dockercoins ``` - See modifications in the code: ```bash git log -p --since "4-JUL-2015" -- . ':!*.yml*' ':!*.html' ``` ] .debug[[swarm/swarmready.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmready.md)] --- ## Which files have been changed since then? - Compose files - HTML file (it contains an embedded contextual tweet) - Dockerfiles (to switch to smaller images) - That's it! -- *We didn't change a single line of code in this app since it was written.* -- *The images that were [built in June 2015]( https://hub.docker.com/r/jpetazzo/dockercoins_worker/tags/) (when the app was written) can still run today ...
... in Swarm mode (distributed across a cluster, with load balancing) ...
... without any modification.* .debug[[swarm/swarmready.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmready.md)] --- ## How did we design our app in the first place? - [Twelve-Factor App](https://12factor.net/) principles - Service discovery using DNS names - Initially implemented as "links" - Then "ambassadors" - And now "services" - Existing apps might require more changes! .debug[[swarm/swarmready.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmready.md)] --- class: btp-manual ## Integration with Compose - We saw how to manually build, tag, and push images to a registry - But ... -- class: btp-manual *"I'm so glad that my deployment relies on ten nautic miles of Shell scripts"* *(No-one, ever)* -- class: btp-manual - Let's see how we can streamline this process! .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/train-of-containers-1.jpg)] --- name: toc-swarm-stacks class: title Swarm Stacks .nav[ [Previous part](#toc-global-scheduling) | [Back to table of contents](#toc-part-2) | [Next part](#toc-cicd-for-docker-and-orchestration) ] .debug[(automatically generated title slide)] --- # Swarm Stacks - Compose is great for local development - It can also be used to manage image lifecycle (i.e. build images and push them to a registry) - Compose files *v2* are great for local development - Compose files *v3* can also be used for production deployments! .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Compose file version 3 (New in Docker Engine 1.13) - Almost identical to version 2 - Can be directly used by a Swarm cluster through `docker stack ...` commands - Introduces a `deploy` section to pass Swarm-specific parameters - Resource limits are moved to this `deploy` section - See [here](https://github.com/docker/docker.github.io/blob/master/compose/compose-file/compose-versioning.md#upgrading) for the complete list of changes - Supersedes *Distributed Application Bundles* (JSON payload describing an application; could be generated from a Compose file) .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Our first stack We need a registry to move images around. Without a stack file, it would be deployed with the following command: ```bash docker service create --publish 5000:5000 registry ``` Now, we are going to deploy it with the following stack file: ```yaml version: "3" services: registry: image: registry ports: - "5000:5000" ``` .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Checking our stack files - All the stack files that we will use are in the `stacks` directory .lab[ - Go to the `stacks` directory: ```bash cd ~/container.training/stacks ``` - Check `registry.yml`: ```bash cat registry.yml ``` ] .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Deploying our first stack - All stack manipulation commands start with `docker stack` - Under the hood, they map to `docker service` commands - Stacks have a *name* (which also serves as a namespace) - Stacks are specified with the aforementioned Compose file format version 3 .lab[ - Deploy our local registry: ```bash docker stack deploy --compose-file registry.yml registry ``` ] .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Inspecting stacks - `docker stack ps` shows the detailed state of all services of a stack .lab[ - Check that our registry is running correctly: ```bash docker stack ps registry ``` - Confirm that we get the same output with the following command: ```bash docker service ps registry_registry ``` ] .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- class: btp-manual ## Specifics of stack deployment Our registry is not *exactly* identical to the one deployed with `docker service create`! - Each stack gets its own overlay network - Services of the stack are connected to this network
(unless specified differently in the Compose file) - Services get network aliases matching their name in the Compose file
(just like when Compose brings up an app specified in a v2 file) - Services are explicitly named `
_
` - Services and tasks also get an internal label indicating which stack they belong to .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- class: btp-auto ## Testing our local registry - Connecting to port 5000 *on any node of the cluster* routes us to the registry - Therefore, we can use `localhost:5000` or `127.0.0.1:5000` as our registry .lab[ - Issue the following API request to the registry: ```bash curl 127.0.0.1:5000/v2/_catalog ``` ] It should return: ```json {"repositories":[]} ``` If that doesn't work, retry a few times; perhaps the container is still starting. .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- class: btp-auto ## Pushing an image to our local registry - We can retag a small image, and push it to the registry .lab[ - Make sure we have the busybox image, and retag it: ```bash docker pull busybox docker tag busybox 127.0.0.1:5000/busybox ``` - Push it: ```bash docker push 127.0.0.1:5000/busybox ``` ] .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- class: btp-auto ## Checking what's on our local registry - The registry API has endpoints to query what's there .lab[ - Ensure that our busybox image is now in the local registry: ```bash curl http://127.0.0.1:5000/v2/_catalog ``` ] The curl command should now output: ```json "repositories":["busybox"]} ``` .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Building and pushing stack services - When using Compose file version 2 and above, you can specify *both* `build` and `image` - When both keys are present: - Compose does "business as usual" (uses `build`) - but the resulting image is named as indicated by the `image` key
(instead of `
_
:latest`) - it can be pushed to a registry with `docker-compose push` - Example: ```yaml webfront: build: www image: myregistry.company.net:5000/webfront ``` .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Using Compose to build and push images .lab[ - Try it: ```bash docker-compose -f dockercoins.yml build docker-compose -f dockercoins.yml push ``` ] Let's have a look at the `dockercoins.yml` file while this is building and pushing. .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ```yaml version: "3" services: rng: build: dockercoins/rng image: ${REGISTRY-127.0.0.1:5000}/rng:${TAG-latest} deploy: mode: global ... redis: image: redis ... worker: build: dockercoins/worker image: ${REGISTRY-127.0.0.1:5000}/worker:${TAG-latest} ... deploy: replicas: 10 ``` .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Deploying the application - Now that the images are on the registry, we can deploy our application stack .lab[ - Create the application stack: ```bash docker stack deploy --compose-file dockercoins.yml dockercoins ``` ] We can now connect to any of our nodes on port 8000, and we will see the familiar hashing speed graph. .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Maintaining multiple environments There are many ways to handle variations between environments. - Compose loads `docker-compose.yml` and (if it exists) `docker-compose.override.yml` - Compose can load alternate file(s) by setting the `-f` flag or the `COMPOSE_FILE` environment variable - Compose files can *extend* other Compose files, selectively including services: ```yaml web: extends: file: common-services.yml service: webapp ``` See [this documentation page](https://docs.docker.com/compose/extends/) for more details about these techniques. .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- class: extra-details ## Good to know ... - Compose file version 3 adds the `deploy` section - Further versions (3.1, ...) add more features (secrets, configs ...) - You can re-run `docker stack deploy` to update a stack - You can make manual changes with `docker service update` ... - ... But they will be wiped out each time you `docker stack deploy` (That's the intended behavior, when one thinks about it!) - `extends` doesn't work with `docker stack deploy` (But you can use `docker-compose config` to "flatten" your configuration) .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- ## Summary - We've seen how to set up a Swarm - We've used it to host our own registry - We've built our app container images - We've used the registry to host those images - We've deployed and scaled our application - We've seen how to use Compose to streamline deployments - Awesome job, team! .debug[[swarm/stacks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stacks.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/train-of-containers-2.jpg)] --- name: toc-cicd-for-docker-and-orchestration class: title CI/CD for Docker and orchestration .nav[ [Previous part](#toc-swarm-stacks) | [Back to table of contents](#toc-part-2) | [Next part](#toc-breaking-into-an-overlay-network) ] .debug[(automatically generated title slide)] --- name: cicd # CI/CD for Docker and orchestration A quick note about continuous integration and deployment - This lab won't have you building out CI/CD pipelines - We're cheating a bit by building images on server hosts and not in CI tool - Docker and orchestration works with all the CI and deployment tools .debug[[swarm/cicd.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/cicd.md)] --- ## CI/CD general process - Have your CI build your images, run tests *in them*, then push to registry - If you security scan, do it then on your images after tests but before push - Optionally, have CI do continuous deployment if build/test/push is successful - CD tool would SSH into nodes, or use docker cli against remote engine - If supported, it could use docker engine TCP API (swarm API is built-in) - Docker KBase [Development Pipeline Best Practices](https://success.docker.com/article/dev-pipeline) - Docker KBase [Continuous Integration with Docker Hub](https://success.docker.com/article/continuous-integration-with-docker-hub) - Docker KBase [Building a Docker Secure Supply Chain](https://success.docker.com/article/secure-supply-chain) .debug[[swarm/cicd.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/cicd.md)] --- class: pic ![CI-CD with Docker](images/ci-cd-with-docker.png) .debug[[swarm/cicd.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/cicd.md)] --- name: part-2 class: title, self-paced Part 2 .debug[[swarm-selfpaced.yml](https://github.com/jpetazzo/container.training/tree/main/slides/swarm-selfpaced.yml)] --- class: self-paced ## Before we continue ... The following section assumes that you have a 5-nodes Swarm cluster. If you come here from a previous tutorial and still have your cluster: great! Otherwise: check [part 1](#part-1) to learn how to set up your own cluster. We pick up exactly where we left you, so we assume that you have: - a Swarm cluster with at least 3 nodes, - a self-hosted registry, - DockerCoins up and running. The next slide has a cheat sheet if you need to set that up in a pinch. .debug[[swarm/operatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/operatingswarm.md)] --- class: self-paced ## Catching up Assuming you have 5 nodes provided by [Play-With-Docker](https://www.play-with-docker/), do this from `node1`: ```bash docker swarm init --advertise-addr eth0 TOKEN=$(docker swarm join-token -q manager) for N in $(seq 2 5); do DOCKER_HOST=tcp://node$N:2375 docker swarm join --token $TOKEN node1:2377 done git clone https://github.com/jpetazzo/container.training cd container.training/stacks docker stack deploy --compose-file registry.yml registry docker-compose -f dockercoins.yml build docker-compose -f dockercoins.yml push docker stack deploy --compose-file dockercoins.yml dockercoins ``` You should now be able to connect to port 8000 and see the DockerCoins web UI. .debug[[swarm/operatingswarm.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/operatingswarm.md)] --- class: extra-details ## Troubleshooting overlay networks - We want to run tools like `ab` or `httping` on the internal network -- class: extra-details - Ah, if only we had created our overlay network with the `--attachable` flag ... -- class: extra-details - Oh well, let's use this as an excuse to introduce New Ways To Do Things .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/two-containers-on-a-truck.jpg)] --- name: toc-breaking-into-an-overlay-network class: title Breaking into an overlay network .nav[ [Previous part](#toc-cicd-for-docker-and-orchestration) | [Back to table of contents](#toc-part-3) | [Next part](#toc-securing-overlay-networks) ] .debug[(automatically generated title slide)] --- # Breaking into an overlay network - We will create a dummy placeholder service on our network - Then we will use `docker exec` to run more processes in this container .lab[ - Start a "do nothing" container using our favorite Swiss-Army distro: ```bash docker service create --network dockercoins_default --name debug \ --constraint node.hostname==$HOSTNAME alpine sleep 1000000000 ``` ] The `constraint` makes sure that the container will be created on the local node. .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- ## Entering the debug container - Once our container is started (which should be really fast because the alpine image is small), we can enter it (from any node) .lab[ - Locate the container: ```bash docker ps ``` - Enter it: ```bash docker exec -ti containerID sh ``` ] .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- ## Labels - We can also be fancy and find the ID of the container automatically - SwarmKit places labels on containers .lab[ - Get the ID of the container: ```bash CID=$(docker ps -q --filter label=com.docker.swarm.service.name=debug) ``` - And enter the container: ```bash docker exec -ti $CID sh ``` ] .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- ## Installing our debugging tools - Ideally, you would author your own image, with all your favorite tools, and use it instead of the base `alpine` image - But we can also dynamically install whatever we need .lab[ - Install a few tools: ```bash apk add --update curl apache2-utils drill ``` ] .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- ## Investigating the `rng` service - First, let's check what `rng` resolves to .lab[ - Use drill or nslookup to resolve `rng`: ```bash drill rng ``` ] This give us one IP address. It is not the IP address of a container. It is a virtual IP address (VIP) for the `rng` service. .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- ## Investigating the VIP .lab[ - Try to ping the VIP: ```bash ping -c 3 rng ``` ] It *should* ping. (But this might change in the future.) With Engine 1.12: VIPs respond to ping if a backend is available on the same machine. With Engine 1.13: VIPs respond to ping if a backend is available anywhere. (Again: this might change in the future.) .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- ## What if I don't like VIPs? - Services can be published using two modes: VIP and DNSRR. - With VIP, you get a virtual IP for the service, and a load balancer based on IPVS (By the way, IPVS is totally awesome and if you want to learn more about it in the context of containers, I highly recommend [this talk](https://www.youtube.com/watch?v=oFsJVV1btDU&index=5&list=PLkA60AVN3hh87OoVra6MHf2L4UR9xwJkv) by [@kobolog](https://twitter.com/kobolog) at DC15EU!) - With DNSRR, you get the former behavior (from Engine 1.11), where resolving the service yields the IP addresses of all the containers for this service - You change this with `docker service create --endpoint-mode [VIP|DNSRR]` .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- ## Looking up VIP backends - You can also resolve a special name: `tasks.
` - It will give you the IP addresses of the containers for a given service .lab[ - Obtain the IP addresses of the containers for the `rng` service: ```bash drill tasks.rng ``` ] This should list 5 IP addresses. .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, benchmarking ## Testing and benchmarking our service - We will check that the service is up with `rng`, then benchmark it with `ab` .lab[ - Make a test request to the service: ```bash curl rng ``` - Open another window, and stop the workers, to test in isolation: ```bash docker service update dockercoins_worker --replicas 0 ``` ] Wait until the workers are stopped (check with `docker service ls`) before continuing. .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, benchmarking ## Benchmarking `rng` We will send 50 requests, but with various levels of concurrency. .lab[ - Send 50 requests, with a single sequential client: ```bash ab -c 1 -n 50 http://rng/10 ``` - Send 50 requests, with fifty parallel clients: ```bash ab -c 50 -n 50 http://rng/10 ``` ] .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, benchmarking ## Benchmark results for `rng` - When serving requests sequentially, they each take 100ms - In the parallel scenario, the latency increased dramatically: - What about `hasher`? .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, benchmarking ## Benchmarking `hasher` We will do the same tests for `hasher`. The command is slightly more complex, since we need to post random data. First, we need to put the POST payload in a temporary file. .lab[ - Generate 10 bytes of random data: ```bash curl http://rng/10 >/tmp/random ``` ] .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, benchmarking ## Benchmarking `hasher` Once again, we will send 50 requests, with different levels of concurrency. .lab[ - Send 50 requests with a sequential client: ```bash ab -c 1 -n 50 -T application/octet-stream -p /tmp/random http://hasher/ ``` - Send 50 requests with 50 parallel clients: ```bash ab -c 50 -n 50 -T application/octet-stream -p /tmp/random http://hasher/ ``` ] .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, benchmarking ## Benchmark results for `hasher` - The sequential benchmarks takes ~5 seconds to complete - The parallel benchmark takes less than 1 second to complete - In both cases, each request takes a bit more than 100ms to complete - Requests are a bit slower in the parallel benchmark - It looks like `hasher` is better equipped to deal with concurrency than `rng` .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, title, benchmarking Why? .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, benchmarking ## Why does everything take (at least) 100ms? `rng` code: ![RNG code screenshot](images/delay-rng.png) `hasher` code: ![HASHER code screenshot](images/delay-hasher.png) .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, title, benchmarking But ... WHY?!? .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, benchmarking ## Why did we sprinkle the code with sleeps? - Deterministic performance
(regardless of instance speed, CPUs, I/O...) - Actual code sleeps all the time anyway - When your code makes a remote API call: - it sends a request; - it sleeps until it gets the response; - it processes the response. .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details, in-person, benchmarking ## Why do `rng` and `hasher` behave differently? ![Equations on a blackboard](images/equations.png) (Synchronous vs. asynchronous event processing) .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: extra-details ## Global scheduling → global debugging - Traditional approach: - log into a node - install our Swiss Army Knife (if necessary) - troubleshoot things - Proposed alternative: - put our Swiss Army Knife in a container (e.g. [nicolaka/netshoot](https://hub.docker.com/r/nicolaka/netshoot/)) - run tests from multiple locations at the same time (This becomes very practical with the `docker service log` command, available since 17.05.) .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- ## More about overlay networks .blackbelt[DC17US: Deep Dive in Docker Overlay Networks ([video](https://www.youtube.com/watch?v=b3XDl0YsVsg&index=1&list=PLkA60AVN3hh-biQ6SCtBJ-WVTyBmmYho8))] .blackbelt[DC17EU: Deeper Dive in Docker Overlay Networks ([video](https://dockercon.docker.com/watch/XkRRA7Etsznv7uAk1UKsri))] .debug[[swarm/netshoot.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/netshoot.md)] --- class: nbt, extra-details ## Measuring cluster-wide network conditions - Since we have built-in, cluster-wide discovery, it's relatively straightforward to monitor the whole cluster automatically - [Alexandros Mavrogiannis](https://github.com/alexmavr) wrote [Swarm NBT](https://github.com/alexmavr/swarm-nbt), a tool doing exactly that! .lab[ - Start Swarm NBT: ```bash docker run --rm -v inventory:/inventory \ -v /var/run/docker.sock:/var/run/docker.sock \ alexmavr/swarm-nbt start ``` ] Note: in this mode, Swarm NBT connects to the Docker API socket, and issues additional API requests to start all the components it needs. .debug[[swarm/swarmnbt.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmnbt.md)] --- class: nbt, extra-details ## Viewing network conditions with Prometheus - Swarm NBT relies on Prometheus to scrape and store data - We can directly consume the Prometheus endpoint to view telemetry data .lab[ - Point your browser to any Swarm node, on port 9090 (If you're using Play-With-Docker, click on the (9090) badge) - In the drop-down, select `icmp_rtt_gauge_seconds` - Click on "Graph" ] You are now seeing ICMP latency across your cluster. .debug[[swarm/swarmnbt.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmnbt.md)] --- class: nbt, in-person, extra-details ## Viewing network conditions with Grafana - If you are using a "real" cluster (not Play-With-Docker) you can use Grafana .lab[ - Start Grafana with `docker service create -p 3000:3000 grafana` - Point your browser to Grafana, on port 3000 on any Swarm node - Login with username `admin` and password `admin` - Click on the top-left menu and browse to Data Sources - Create a prometheus datasource with any name - Point it to http://any-node-IP:9090 - Set access to "direct" and leave credentials blank - Click on the top-left menu, highlight "Dashboards" and select the "Import" option - Copy-paste [this JSON payload]( https://raw.githubusercontent.com/alexmavr/swarm-nbt/master/grafana.json), then use the Prometheus Data Source defined before - Poke around the dashboard that magically appeared! ] .debug[[swarm/swarmnbt.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmnbt.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/wall-of-containers.jpeg)] --- name: toc-securing-overlay-networks class: title Securing overlay networks .nav[ [Previous part](#toc-breaking-into-an-overlay-network) | [Back to table of contents](#toc-part-3) | [Next part](#toc-updating-services) ] .debug[(automatically generated title slide)] --- # Securing overlay networks - By default, overlay networks are using plain VXLAN encapsulation (~Ethernet over UDP, using SwarmKit's control plane for ARP resolution) - Encryption can be enabled on a per-network basis (It will use IPSEC encryption provided by the kernel, leveraging hardware acceleration) - This is only for the `overlay` driver (Other drivers/plugins will use different mechanisms) .debug[[swarm/ipsec.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ipsec.md)] --- ## Creating two networks: encrypted and not - Let's create two networks for testing purposes .lab[ - Create an "insecure" network: ```bash docker network create insecure --driver overlay --attachable ``` - Create a "secure" network: ```bash docker network create secure --opt encrypted --driver overlay --attachable ``` ] .warning[Make sure that you don't typo that option; errors are silently ignored!] .debug[[swarm/ipsec.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ipsec.md)] --- ## Deploying a dual-homed web server - Let's use good old NGINX - We will attach it to both networks - We will use a placement constraint to make sure that it is on a different node .lab[ - Create a web server running somewhere else: ```bash docker service create --name web \ --network secure --network insecure \ --constraint node.hostname!=node1 \ nginx ``` ] .debug[[swarm/ipsec.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ipsec.md)] --- ## Sniff HTTP traffic - We will use `ngrep`, which allows to grep for network traffic - We will run it in a container, using host networking to access the host's interfaces .lab[ - Sniff network traffic and display all packets containing "HTTP": ```bash docker run --net host nicolaka/netshoot ngrep -tpd eth0 HTTP ``` ] -- Seeing tons of HTTP request? Shutdown your DockerCoins workers: ```bash docker service update dockercoins_worker --replicas=0 ``` .debug[[swarm/ipsec.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ipsec.md)] --- ## Check that we are, indeed, sniffing traffic - Let's see if we can intercept our traffic with Google! .lab[ - Open a new terminal - Issue an HTTP request to Google (or anything you like): ```bash curl google.com ``` ] The ngrep container will display one `#` per packet traversing the network interface. When you do the `curl`, you should see the HTTP request in clear text in the output. .debug[[swarm/ipsec.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ipsec.md)] --- class: extra-details ## If you are using Play-With-Docker, Vagrant, etc. - You will probably have *two* network interfaces - One interface will be used for outbound traffic (to Google) - The other one will be used for internode traffic - You might have to adapt/relaunch the `ngrep` command to specify the right one! .debug[[swarm/ipsec.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ipsec.md)] --- ## Try to sniff traffic across overlay networks - We will run `curl web` through both secure and insecure networks .lab[ - Access the web server through the insecure network: ```bash docker run --rm --net insecure nicolaka/netshoot curl web ``` - Now do the same through the secure network: ```bash docker run --rm --net secure nicolaka/netshoot curl web ``` ] When you run the first command, you will see HTTP fragments.
However, when you run the second one, only `#` will show up. .debug[[swarm/ipsec.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/ipsec.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/catene-de-conteneurs.jpg)] --- name: toc-updating-services class: title Updating services .nav[ [Previous part](#toc-securing-overlay-networks) | [Back to table of contents](#toc-part-3) | [Next part](#toc-rolling-updates) ] .debug[(automatically generated title slide)] --- # Updating services - We want to make changes to the web UI - The process is as follows: - edit code - build new image - ship new image - run new image .debug[[swarm/updatingservices.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/updatingservices.md)] --- ## Updating a single service with `service update` - To update a single service, we could do the following: ```bash export REGISTRY=127.0.0.1:5000 export TAG=v0.2 IMAGE=$REGISTRY/dockercoins_webui:$TAG docker build -t $IMAGE webui/ docker push $IMAGE docker service update dockercoins_webui --image $IMAGE ``` - Make sure to tag properly your images: update the `TAG` at each iteration (When you check which images are running, you want these tags to be uniquely identifiable) .debug[[swarm/updatingservices.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/updatingservices.md)] --- ## Updating services with `stack deploy` - With the Compose integration, all we have to do is: ```bash export TAG=v0.2 docker-compose -f composefile.yml build docker-compose -f composefile.yml push docker stack deploy -c composefile.yml nameofstack ``` -- - That's exactly what we used earlier to deploy the app - We don't need to learn new commands! - It will diff each service and only update ones that changed .debug[[swarm/updatingservices.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/updatingservices.md)] --- ## Changing the code - Let's make the numbers on the Y axis bigger! .lab[ - Update the size of text on our webui: ```bash sed -i "s/15px/50px/" dockercoins/webui/files/index.html ``` ] .debug[[swarm/updatingservices.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/updatingservices.md)] --- ## Build, ship, and run our changes - Four steps: 1. Set (and export!) the `TAG` environment variable 2. `docker-compose build` 3. `docker-compose push` 4. `docker stack deploy` .lab[ - Build, ship, and run: ```bash export TAG=v0.2 docker-compose -f dockercoins.yml build docker-compose -f dockercoins.yml push docker stack deploy -c dockercoins.yml dockercoins ``` ] - Because we're tagging all images in this demo v0.2, deploy will update all apps, FYI .debug[[swarm/updatingservices.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/updatingservices.md)] --- ## Viewing our changes - Wait at least 10 seconds (for the new version to be deployed) - Then reload the web UI - Or just mash "reload" frantically - ... Eventually the legend on the left will be bigger! .debug[[swarm/updatingservices.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/updatingservices.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/Container-Ship-Freighter-Navigation-Elbe-Romance-1782991.jpg)] --- name: toc-rolling-updates class: title Rolling updates .nav[ [Previous part](#toc-updating-services) | [Back to table of contents](#toc-part-3) | [Next part](#toc-health-checks-and-auto-rollbacks) ] .debug[(automatically generated title slide)] --- # Rolling updates - Let's force an update on hasher to watch it update .lab[ - First lets scale up hasher to 7 replicas: ```bash docker service scale dockercoins_hasher=7 ``` - Force a rolling update (replace containers) to different image: ```bash docker service update --image 127.0.0.1:5000/hasher:v0.1 dockercoins_hasher ``` ] - You can run `docker events` in a separate `node1` shell to see Swarm actions - You can use `--force` to replace containers without a config change .debug[[swarm/rollingupdates.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/rollingupdates.md)] --- ## Changing the upgrade policy - We can change many options on how updates happen .lab[ - Change the parallelism to 2, and the max failed container updates to 25%: ```bash docker service update --update-parallelism 2 \ --update-max-failure-ratio .25 dockercoins_hasher ``` ] - No containers were replaced, this is called a "no op" change - Service metadata-only changes don't require orchestrator operations .debug[[swarm/rollingupdates.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/rollingupdates.md)] --- ## Changing the policy in the Compose file - The policy can also be updated in the Compose file - This is done by adding an `update_config` key under the `deploy` key: ```yaml deploy: replicas: 10 update_config: parallelism: 2 delay: 10s ``` .debug[[swarm/rollingupdates.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/rollingupdates.md)] --- ## Rolling back - At any time (e.g. before the upgrade is complete), we can rollback: - by editing the Compose file and redeploying - by using the special `--rollback` flag with `service update` - by using `docker service rollback` .lab[ - Try to rollback the webui service: ```bash docker service rollback dockercoins_webui ``` ] What happens with the web UI graph? .debug[[swarm/rollingupdates.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/rollingupdates.md)] --- ## The fine print with rollback - Rollback reverts to the previous service definition - see `PreviousSpec` in `docker service inspect
` - If we visualize successive updates as a stack: - it doesn't "pop" the latest update - it "pushes" a copy of the previous update on top - ergo, rolling back twice does nothing - "Service definition" includes rollout cadence - Each `docker service update` command = a new service definition .debug[[swarm/rollingupdates.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/rollingupdates.md)] --- class: extra-details ## Timeline of an upgrade - SwarmKit will upgrade N instances at a time
(following the `update-parallelism` parameter) - New tasks are created, and their desired state is set to `Ready`
.small[(this pulls the image if necessary, ensures resource availability, creates the container ... without starting it)] - If the new tasks fail to get to `Ready` state, go back to the previous step
.small[(SwarmKit will try again and again, until the situation is addressed or desired state is updated)] - When the new tasks are `Ready`, it sets the old tasks desired state to `Shutdown` - When the old tasks are `Shutdown`, it starts the new tasks - Then it waits for the `update-delay`, and continues with the next batch of instances .debug[[swarm/rollingupdates.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/rollingupdates.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/ShippingContainerSFBay.jpg)] --- name: toc-health-checks-and-auto-rollbacks class: title Health checks and auto-rollbacks .nav[ [Previous part](#toc-rolling-updates) | [Back to table of contents](#toc-part-3) | [Next part](#toc-swarmkit-debugging-tools) ] .debug[(automatically generated title slide)] --- name: healthchecks # Health checks and auto-rollbacks (New in Docker Engine 1.12) - Commands that are executed on regular intervals in a container - Must return 0 or 1 to indicate "all is good" or "something's wrong" - Must execute quickly (timeouts = failures) - Example: ```bash curl -f http://localhost/_ping || false ``` - the `-f` flag ensures that `curl` returns non-zero for 404 and similar errors - `|| false` ensures that any non-zero exit status gets mapped to 1 - `curl` must be installed in the container that is being checked .debug[[swarm/healthchecks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/healthchecks.md)] --- ## Defining health checks - In a Dockerfile, with the [HEALTHCHECK](https://docs.docker.com/engine/reference/builder/#healthcheck) instruction ``` HEALTHCHECK --interval=1s --timeout=3s CMD curl -f http://localhost/ || false ``` - From the command line, when running containers or services ``` docker run --health-cmd "curl -f http://localhost/ || false" ... docker service create --health-cmd "curl -f http://localhost/ || false" ... ``` - In Compose files, with a per-service [healthcheck](https://docs.docker.com/compose/compose-file/#healthcheck) section ```yaml www: image: hellowebapp healthcheck: test: "curl -f https://localhost/ || false" timeout: 3s ``` .debug[[swarm/healthchecks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/healthchecks.md)] --- ## Using health checks - With `docker run`, health checks are purely informative - `docker ps` shows health status - `docker inspect` has extra details (including health check command output) - With `docker service`: - unhealthy tasks are terminated (i.e. the service is restarted) - failed deployments can be rolled back automatically
(by setting *at least* the flag `--update-failure-action rollback`) .debug[[swarm/healthchecks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/healthchecks.md)] --- ## Enabling health checks and auto-rollbacks Here is a comprehensive example using the CLI: .small[ ```bash docker service update \ --update-delay 5s \ --update-failure-action rollback \ --update-max-failure-ratio .25 \ --update-monitor 5s \ --update-parallelism 1 \ --rollback-delay 5s \ --rollback-failure-action pause \ --rollback-max-failure-ratio .5 \ --rollback-monitor 5s \ --rollback-parallelism 0 \ --health-cmd "curl -f http://localhost/ || exit 1" \ --health-interval 2s \ --health-retries 1 \ --image yourimage:newversion yourservice ``` ] .debug[[swarm/healthchecks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/healthchecks.md)] --- ## Implementing auto-rollback in practice We will use the following Compose file (`stacks/dockercoins+healthcheck.yml`): ```yaml ... hasher: build: dockercoins/hasher image: ${REGISTRY-127.0.0.1:5000}/hasher:${TAG-latest} healthcheck: test: curl -f http://localhost/ || exit 1 deploy: replicas: 7 update_config: delay: 5s failure_action: rollback max_failure_ratio: .5 monitor: 5s parallelism: 1 ... ``` .debug[[swarm/healthchecks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/healthchecks.md)] --- ## Enabling auto-rollback in dockercoins We need to update our services with a healthcheck. .lab[ - Go to the `stacks` directory: ```bash cd ~/container.training/stacks ``` - Deploy the updated stack with healthchecks built-in: ```bash docker stack deploy --compose-file dockercoins+healthcheck.yml dockercoins ``` ] .debug[[swarm/healthchecks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/healthchecks.md)] --- ## Visualizing an automated rollback - Here's a good example of why healthchecks are necessary - This breaking change will prevent the app from listening on the correct port - The container still runs fine, it just won't accept connections on port 80 .lab[ - Change the HTTP listening port: ```bash sed -i "s/80/81/" dockercoins/hasher/hasher.rb ``` - Build, ship, and run the new image: ```bash export TAG=v0.3 docker-compose -f dockercoins+healthcheck.yml build docker-compose -f dockercoins+healthcheck.yml push docker service update --image=127.0.0.1:5000/hasher:$TAG dockercoins_hasher ``` ] .debug[[swarm/healthchecks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/healthchecks.md)] --- ## CLI flags for health checks and rollbacks .small[ ``` --health-cmd string Command to run to check health --health-interval duration Time between running the check (ms|s|m|h) --health-retries int Consecutive failures needed to report unhealthy --health-start-period duration Start period for the container to initialize before counting retries towards unstable (ms|s|m|h) --health-timeout duration Maximum time to allow one check to run (ms|s|m|h) --no-healthcheck Disable any container-specified HEALTHCHECK --restart-condition string Restart when condition is met ("none"|"on-failure"|"any") --restart-delay duration Delay between restart attempts (ns|us|ms|s|m|h) --restart-max-attempts uint Maximum number of restarts before giving up --restart-window duration Window used to evaluate the restart policy (ns|us|ms|s|m|h) --rollback Rollback to previous specification --rollback-delay duration Delay between task rollbacks (ns|us|ms|s|m|h) --rollback-failure-action string Action on rollback failure ("pause"|"continue") --rollback-max-failure-ratio float Failure rate to tolerate during a rollback --rollback-monitor duration Duration after each task rollback to monitor for failure (ns|us|ms|s|m|h) --rollback-order string Rollback order ("start-first"|"stop-first") --rollback-parallelism uint Maximum number of tasks rolled back simultaneously (0 to roll back all at once) --update-delay duration Delay between updates (ns|us|ms|s|m|h) --update-failure-action string Action on update failure ("pause"|"continue"|"rollback") --update-max-failure-ratio float Failure rate to tolerate during an update --update-monitor duration Duration after each task update to monitor for failure (ns|us|ms|s|m|h) --update-order string Update order ("start-first"|"stop-first") --update-parallelism uint Maximum number of tasks updated simultaneously (0 to update all at once) ``` ] .debug[[swarm/healthchecks.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/healthchecks.md)] --- ## Getting task information for a given node - You can see all the tasks assigned to a node with `docker node ps` - It shows the *desired state* and *current state* of each task - `docker node ps` shows info about the current node - `docker node ps
` shows info for another node - `docker node ps -f
` allows to select which tasks to show ```bash # Show only tasks that are supposed to be running docker node ps -f desired-state=running # Show only tasks whose name contains the string "front" docker node ps -f name=front ``` .debug[[swarm/nodeinfo.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/nodeinfo.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/aerial-view-of-containers.jpg)] --- name: toc-swarmkit-debugging-tools class: title SwarmKit debugging tools .nav[ [Previous part](#toc-health-checks-and-auto-rollbacks) | [Back to table of contents](#toc-part-3) | [Next part](#toc-secrets-management-and-encryption-at-rest) ] .debug[(automatically generated title slide)] --- # SwarmKit debugging tools - The SwarmKit repository comes with debugging tools - They are *low level* tools; not for general use - We are going to see two of these tools: - `swarmctl`, to communicate directly with the SwarmKit API - `swarm-rafttool`, to inspect the content of the Raft log .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- ## Building the SwarmKit tools - We are going to install a Go compiler, then download SwarmKit source and build it .lab[ - Download, compile, and install SwarmKit with this one-liner: ```bash docker run -v /usr/local/bin:/go/bin golang \ go get `-v` github.com/docker/swarmkit/... ``` ] Remove `-v` if you don't like verbose things. Shameless promo: for more Go and Docker love, check [this blog post](https://jpetazzo.github.io/2016/09/09/go-docker/)! Note: in the unfortunate event of SwarmKit *master* branch being broken, the build might fail. In that case, just skip the Swarm tools section. .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- ## Getting cluster-wide task information - The Docker API doesn't expose this directly (yet) - But the SwarmKit API does - We are going to query it with `swarmctl` - `swarmctl` is an example program showing how to interact with the SwarmKit API .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- ## Using `swarmctl` - The Docker Engine places the SwarmKit control socket in a special path - You need root privileges to access it .lab[ - If you are using Play-With-Docker, set the following alias: ```bash alias swarmctl='/lib/ld-musl-x86_64.so.1 /usr/local/bin/swarmctl \ --socket /var/run/docker/swarm/control.sock' ``` - Otherwise, set the following alias: ```bash alias swarmctl='sudo swarmctl \ --socket /var/run/docker/swarm/control.sock' ``` ] .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- ## `swarmctl` in action - Let's review a few useful `swarmctl` commands .lab[ - List cluster nodes (that's equivalent to `docker node ls`): ```bash swarmctl node ls ``` - View all tasks across all services: ```bash swarmctl task ls ``` ] .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- ## `swarmctl` notes - SwarmKit is vendored into the Docker Engine - If you want to use `swarmctl`, you need the exact version of SwarmKit that was used in your Docker Engine - Otherwise, you might get some errors like: ``` Error: grpc: failed to unmarshal the received message proto: wrong wireType = 0 ``` - With Docker 1.12, the control socket was in `/var/lib/docker/swarm/control.sock` .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- ## `swarm-rafttool` - SwarmKit stores all its important data in a distributed log using the Raft protocol (This log is also simply called the "Raft log") - You can decode that log with `swarm-rafttool` - This is a great tool to understand how SwarmKit works - It can also be used in forensics or troubleshooting (But consider it as a *very low level* tool!) .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- ## The powers of `swarm-rafttool` With `swarm-rafttool`, you can: - view the latest snapshot of the cluster state; - view the Raft log (i.e. changes to the cluster state); - view specific objects from the log or snapshot; - decrypt the Raft data (to analyze it with other tools). It *cannot* work on live files, so you must stop Docker or make a copy first. .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- ## Using `swarm-rafttool` - First, let's make a copy of the current Swarm data .lab[ - If you are using Play-With-Docker, the Docker data directory is `/graph`: ```bash cp -r /graph/swarm /swarmdata ``` - Otherwise, it is in the default `/var/lib/docker`: ```bash sudo cp -r /var/lib/docker/swarm /swarmdata ``` ] .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- ## Dumping the Raft log - We have to indicate the path holding the Swarm data (Otherwise `swarm-rafttool` will try to use the live data, and complain that it's locked!) .lab[ - If you are using Play-With-Docker, you must use the musl linker: ```bash /lib/ld-musl-x86_64.so.1 /usr/local/bin/swarm-rafttool -d /swarmdata/ dump-wal ``` - Otherwise, you don't need the musl linker but you need to get root: ```bash sudo swarm-rafttool -d /swarmdata/ dump-wal ``` ] Reminder: this is a very low-level tool, requiring a knowledge of SwarmKit's internals! .debug[[swarm/swarmtools.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/swarmtools.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/blue-containers.jpg)] --- name: toc-secrets-management-and-encryption-at-rest class: title Secrets management and encryption at rest .nav[ [Previous part](#toc-swarmkit-debugging-tools) | [Back to table of contents](#toc-part-4) | [Next part](#toc-least-privilege-model) ] .debug[(automatically generated title slide)] --- # Secrets management and encryption at rest (New in Docker Engine 1.13) - Secrets management = selectively and securely bring secrets to services - Encryption at rest = protect against storage theft or prying - Remember: - control plane is authenticated through mutual TLS, certs rotated every 90 days - control plane is encrypted with AES-GCM, keys rotated every 12 hours - data plane is not encrypted by default (for performance reasons),
but we saw earlier how to enable that with a single flag .debug[[swarm/security.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/security.md)] --- class: secrets ## Secret management - Docker has a "secret safe" (secure key→value store) - You can create as many secrets as you like - You can associate secrets to services - Secrets are exposed as plain text files, but kept in memory only (using `tmpfs`) - Secrets are immutable (at least in Engine 1.13) - Secrets have a max size of 500 KB .debug[[swarm/secrets.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/secrets.md)] --- class: secrets ## Creating secrets - Must specify a name for the secret; and the secret itself .lab[ - Assign [one of the four most commonly used passwords](https://www.youtube.com/watch?v=0Jx8Eay5fWQ) to a secret called `hackme`: ```bash echo love | docker secret create hackme - ``` ] If the secret is in a file, you can simply pass the path to the file. (The special path `-` indicates to read from the standard input.) .debug[[swarm/secrets.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/secrets.md)] --- class: secrets ## Creating better secrets - Picking lousy passwords always leads to security breaches .lab[ - Let's craft a better password, and assign it to another secret: ```bash base64 /dev/urandom | head -c16 | docker secret create arewesecureyet - ``` ] Note: in the latter case, we don't even know the secret at this point. But Swarm does. .debug[[swarm/secrets.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/secrets.md)] --- class: secrets ## Using secrets - Secrets must be handed explicitly to services .lab[ - Create a dummy service with both secrets: ```bash docker service create \ --secret hackme --secret arewesecureyet \ --name dummyservice \ --constraint node.hostname==$HOSTNAME \ alpine sleep 1000000000 ``` ] We constrain the container to be on the local node for convenience.
(We are going to use `docker exec` in just a moment!) .debug[[swarm/secrets.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/secrets.md)] --- class: secrets ## Accessing secrets - Secrets are materialized on `/run/secrets` (which is an in-memory filesystem) .lab[ - Find the ID of the container for the dummy service: ```bash CID=$(docker ps -q --filter label=com.docker.swarm.service.name=dummyservice) ``` - Enter the container: ```bash docker exec -ti $CID sh ``` - Check the files in `/run/secrets` ] .debug[[swarm/secrets.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/secrets.md)] --- class: secrets ## Rotating secrets - You can't change a secret (Sounds annoying at first; but allows clean rollbacks if a secret update goes wrong) - You can add a secret to a service with `docker service update --secret-add` (This will redeploy the service; it won't add the secret on the fly) - You can remove a secret with `docker service update --secret-rm` - Secrets can be mapped to different names by expressing them with a micro-format: ```bash docker service create --secret source=secretname,target=filename ``` .debug[[swarm/secrets.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/secrets.md)] --- class: secrets ## Changing our insecure password - We want to replace our `hackme` secret with a better one .lab[ - Remove the insecure `hackme` secret: ```bash docker service update dummyservice --secret-rm hackme ``` - Add our better secret instead: ```bash docker service update dummyservice \ --secret-add source=arewesecureyet,target=hackme ``` ] Wait for the service to be fully updated with e.g. `watch docker service ps dummyservice`.
(With Docker Engine 17.10 and later, the CLI will wait for you!) .debug[[swarm/secrets.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/secrets.md)] --- class: secrets ## Checking that our password is now stronger - We will use the power of `docker exec`! .lab[ - Get the ID of the new container: ```bash CID=$(docker ps -q --filter label=com.docker.swarm.service.name=dummyservice) ``` - Check the contents of the secret files: ```bash docker exec $CID grep -r . /run/secrets ``` ] .debug[[swarm/secrets.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/secrets.md)] --- class: secrets ## Secrets in practice - Can be (ab)used to hold whole configuration files if needed - If you intend to rotate secret `foo`, call it `foo.N` instead, and map it to `foo` (N can be a serial, a timestamp...) ```bash docker service create --secret source=foo.N,target=foo ... ``` - You can update (remove+add) a secret in a single command: ```bash docker service update ... --secret-rm foo.M --secret-add source=foo.N,target=foo ``` - For more details and examples, [check the documentation](https://docs.docker.com/engine/swarm/secrets/) .debug[[swarm/secrets.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/secrets.md)] --- ## Encryption at rest - Swarm data is always encrypted - A Swarm cluster can be "locked" - When a cluster is "locked", the encryption key is protected with a passphrase - Starting or restarting a locked manager requires the passphrase - This protects against: - theft (stealing a physical machine, a disk, a backup tape...) - unauthorized access (to e.g. a remote or virtual volume) - some vulnerabilities (like path traversal) .debug[[swarm/encryptionatrest.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/encryptionatrest.md)] --- ## Locking a Swarm cluster - This is achieved through the `docker swarm update` command .lab[ - Lock our cluster: ```bash docker swarm update --autolock=true ``` ] This will display the unlock key. Copy-paste it somewhere safe. .debug[[swarm/encryptionatrest.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/encryptionatrest.md)] --- ## Locked state - If we restart a manager, it will now be locked .lab[ - Restart the local Engine: ```bash sudo systemctl restart docker ``` ] Note: if you are doing the workshop on your own, using nodes that you [provisioned yourself](https://github.com/jpetazzo/container.training/tree/master/prepare-machine) or with [Play-With-Docker](https://play-with-docker.com/), you might have to use a different method to restart the Engine. .debug[[swarm/encryptionatrest.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/encryptionatrest.md)] --- ## Checking that our node is locked - Manager commands (requiring access to crypted data) will fail - Other commands are OK .lab[ - Try a few basic commands: ```bash docker ps docker run alpine echo ♥ docker node ls ``` ] (The last command should fail, and it will tell you how to unlock this node.) .debug[[swarm/encryptionatrest.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/encryptionatrest.md)] --- ## Checking node state in scripts - The state of the node shows up in the output of `docker info` .lab[ - Check the output of `docker info`: ```bash docker info ``` - Can't see it? Too verbose? Grep to the rescue! ```bash docker info | grep ^Swarm ``` ] .debug[[swarm/encryptionatrest.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/encryptionatrest.md)] --- ## Unlocking a node - You will need the secret token that we obtained when enabling auto-lock earlier .lab[ - Unlock the node: ```bash docker swarm unlock ``` - Copy-paste the secret token that we got earlier - Check that manager commands now work correctly: ```bash docker node ls ``` ] .debug[[swarm/encryptionatrest.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/encryptionatrest.md)] --- ## Managing the secret key - If the key is compromised, you can change it and re-encrypt with a new key: ```bash docker swarm unlock-key --rotate ``` - If you lost the key, you can get it as long as you have at least one unlocked node: ```bash docker swarm unlock-key -q ``` Note: if you rotate the key while some nodes are locked, without saving the previous key, those nodes won't be able to rejoin. Note: if somebody steals both your disks and your key, .strike[you're doomed! Doooooomed!]
you can block the compromised node with `docker node demote` and `docker node rm`. .debug[[swarm/encryptionatrest.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/encryptionatrest.md)] --- ## Unlocking the cluster permanently - If you want to remove the secret key, disable auto-lock .lab[ - Permanently unlock the cluster: ```bash docker swarm update --autolock=false ``` ] Note: if some nodes are in locked state at that moment (or if they are offline/restarting while you disabled autolock), they still need the previous unlock key to get back online. For more information about locking, you can check the [upcoming documentation](https://github.com/docker/docker.github.io/pull/694). .debug[[swarm/encryptionatrest.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/encryptionatrest.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/chinook-helicopter-container.jpg)] --- name: toc-least-privilege-model class: title Least privilege model .nav[ [Previous part](#toc-secrets-management-and-encryption-at-rest) | [Back to table of contents](#toc-part-4) | [Next part](#toc-centralized-logging) ] .debug[(automatically generated title slide)] --- # Least privilege model - All the important data is stored in the "Raft log" - Managers nodes have read/write access to this data - Workers nodes have no access to this data - Workers only receive the minimum amount of data that they need: - which services to run - network configuration information for these services - credentials for these services - Compromising a worker node does not give access to the full cluster .debug[[swarm/leastprivilege.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/leastprivilege.md)] --- ## What can I do if I compromise a worker node? - I can enter the containers running on that node - I can access the configuration and credentials used by these containers - I can inspect the network traffic of these containers - I cannot inspect or disrupt the network traffic of other containers (network information is provided by manager nodes; ARP spoofing is not possible) - I cannot infer the topology of the cluster and its number of nodes - I can only learn the IP addresses of the manager nodes .debug[[swarm/leastprivilege.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/leastprivilege.md)] --- ## Guidelines for workload isolation - Define security levels - Define security zones - Put managers in the highest security zone - Enforce workloads of a given security level to run in a given zone - Enforcement can be done with [Authorization Plugins](https://docs.docker.com/engine/extend/plugins_authorization/) .debug[[swarm/leastprivilege.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/leastprivilege.md)] --- ## Learning more about container security .blackbelt[DC17US: Securing Containers, One Patch At A Time ([video](https://www.youtube.com/watch?v=jZSs1RHwcqo&list=PLkA60AVN3hh-biQ6SCtBJ-WVTyBmmYho8&index=4))] .blackbelt[DC17EU: Container-relevant Upstream Kernel Developments ([video](https://dockercon.docker.com/watch/7JQBpvHJwjdW6FKXvMfCK1))] .blackbelt[DC17EU: What Have Syscalls Done for you Lately? ([video](https://dockercon.docker.com/watch/4ZxNyWuwk9JHSxZxgBBi6J))] .debug[[swarm/leastprivilege.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/leastprivilege.md)] --- ## A reminder about *scope* - Out of the box, Docker API access is "all or nothing" - When someone has access to the Docker API, they can access *everything* - If your developers are using the Docker API to deploy on the dev cluster ... ... and the dev cluster is the same as the prod cluster ... ... it means that your devs have access to your production data, passwords, etc. - This can easily be avoided .debug[[swarm/apiscope.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/apiscope.md)] --- ## Fine-grained API access control A few solutions, by increasing order of flexibility: - Use separate clusters for different security perimeters (And different credentials for each cluster) -- - Add an extra layer of abstraction (sudo scripts, hooks, or full-blown PAAS) -- - Enable [authorization plugins] - each API request is vetted by your plugin(s) - by default, the *subject name* in the client TLS certificate is used as user name - example: [user and permission management] in [UCP] [authorization plugins]: https://docs.docker.com/engine/extend/plugins_authorization/ [UCP]: https://docs.docker.com/datacenter/ucp/2.1/guides/ [user and permission management]: https://docs.docker.com/datacenter/ucp/2.1/guides/admin/manage-users/ .debug[[swarm/apiscope.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/apiscope.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/container-cranes.jpg)] --- name: toc-centralized-logging class: title Centralized logging .nav[ [Previous part](#toc-least-privilege-model) | [Back to table of contents](#toc-part-4) | [Next part](#toc-setting-up-elk-to-store-container-logs) ] .debug[(automatically generated title slide)] --- name: logging # Centralized logging - We want to send all our container logs to a central place - If that place could offer a nice web dashboard too, that'd be nice -- - We are going to deploy an ELK stack - It will accept logs over a GELF socket - We will update our services to send logs through the GELF logging driver .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/container-housing.jpg)] --- name: toc-setting-up-elk-to-store-container-logs class: title Setting up ELK to store container logs .nav[ [Previous part](#toc-centralized-logging) | [Back to table of contents](#toc-part-4) | [Next part](#toc-metrics-collection) ] .debug[(automatically generated title slide)] --- # Setting up ELK to store container logs *Important foreword: this is not an "official" or "recommended" setup; it is just an example. We used ELK in this demo because it's a popular setup and we keep being asked about it; but you will have equal success with Fluent or other logging stacks!* What we will do: - Spin up an ELK stack with services - Gaze at the spiffy Kibana web UI - Manually send a few log entries using one-shot containers - Set our containers up to send their logs to Logstash .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## What's in an ELK stack? - ELK is three components: - ElasticSearch (to store and index log entries) - Logstash (to receive log entries from various sources, process them, and forward them to various destinations) - Kibana (to view/search log entries with a nice UI) - The only component that we will configure is Logstash - We will accept log entries using the GELF protocol - Log entries will be stored in ElasticSearch,
and displayed on Logstash's stdout for debugging .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- class: elk-manual ## Setting up ELK - We need three containers: ElasticSearch, Logstash, Kibana - We will place them on a common network, `logging` .lab[ - Create the network: ```bash docker network create --driver overlay logging ``` - Create the ElasticSearch service: ```bash docker service create --network logging --name elasticsearch elasticsearch:2.4 ``` ] .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- class: elk-manual ## Setting up Kibana - Kibana exposes the web UI - Its default port (5601) needs to be published - It needs a tiny bit of configuration: the address of the ElasticSearch service - We don't want Kibana logs to show up in Kibana (it would create clutter)
so we tell Logspout to ignore them .lab[ - Create the Kibana service: ```bash docker service create --network logging --name kibana --publish 5601:5601 \ -e ELASTICSEARCH_URL=http://elasticsearch:9200 kibana:4.6 ``` ] .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- class: elk-manual ## Setting up Logstash - Logstash needs some configuration to listen to GELF messages and send them to ElasticSearch - We could author a custom image bundling this configuration - We can also pass the [configuration](https://github.com/jpetazzo/container.training/blob/master/elk/logstash.conf) on the command line .lab[ - Create the Logstash service: ```bash docker service create --network logging --name logstash -p 12201:12201/udp \ logstash:2.4 -e "$(cat ~/container.training/elk/logstash.conf)" ``` ] .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- class: elk-manual ## Checking Logstash - Before proceeding, let's make sure that Logstash started properly .lab[ - Lookup the node running the Logstash container: ```bash docker service ps logstash ``` - Connect to that node ] .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- class: elk-manual ## View Logstash logs .lab[ - View the logs of the logstash service: ```bash docker service logs logstash --follow ``` ] You should see the heartbeat messages: .small[ ```json { "message" => "ok", "host" => "1a4cfb063d13", "@version" => "1", "@timestamp" => "2016-06-19T00:45:45.273Z" } ``` ] .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- class: elk-auto ## Deploying our ELK cluster - We will use a stack file .lab[ - Build, ship, and run our ELK stack: ```bash docker-compose -f elk.yml build docker-compose -f elk.yml push docker stack deploy -c elk.yml elk ``` ] Note: the *build* and *push* steps are not strictly necessary, but they don't hurt! Let's have a look at the [Compose file]( https://github.com/jpetazzo/container.training/blob/master/stacks/elk.yml). .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- class: elk-auto ## Checking that our ELK stack works correctly - Let's view the logs of logstash (Who logs the loggers?) .lab[ - Stream logstash's logs: ```bash docker service logs --follow --tail 1 elk_logstash ``` ] You should see the heartbeat messages: .small[ ```json { "message" => "ok", "host" => "1a4cfb063d13", "@version" => "1", "@timestamp" => "2016-06-19T00:45:45.273Z" } ``` ] .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## Testing the GELF receiver - In a new window, we will generate a logging message - We will use a one-off container, and Docker's GELF logging driver .lab[ - Send a test message: ```bash docker run --log-driver gelf --log-opt gelf-address=udp://127.0.0.1:12201 \ --rm alpine echo hello ``` ] The test message should show up in the logstash container logs. .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## Sending logs from a service - We were sending from a "classic" container so far; let's send logs from a service instead - We're lucky: the parameters (`--log-driver` and `--log-opt`) are exactly the same! .lab[ - Send a test message: ```bash docker service create \ --log-driver gelf --log-opt gelf-address=udp://127.0.0.1:12201 \ alpine echo hello ``` ] The test message should show up as well in the logstash container logs. -- In fact, *multiple messages will show up, and continue to show up every few seconds!* .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## Restart conditions - By default, if a container exits (or is killed with `docker kill`, or runs out of memory ...), the Swarm will restart it (possibly on a different machine) - This behavior can be changed by setting the *restart condition* parameter .lab[ - Change the restart condition so that Swarm doesn't try to restart our container forever: ```bash docker service update `xxx` --restart-condition none ``` ] Available restart conditions are `none`, `any`, and `on-error`. You can also set `--restart-delay`, `--restart-max-attempts`, and `--restart-window`. .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## Connect to Kibana - The Kibana web UI is exposed on cluster port 5601 .lab[ - Connect to port 5601 of your cluster - if you're using Play-With-Docker, click on the (5601) badge above the terminal - otherwise, open http://(any-node-address):5601/ with your browser ] .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## "Configuring" Kibana - If you see a status page with a yellow item, wait a minute and reload (Kibana is probably still initializing) - Kibana should offer you to "Configure an index pattern":
in the "Time-field name" drop down, select "@timestamp", and hit the "Create" button - Then: - click "Discover" (in the top-left corner) - click "Last 15 minutes" (in the top-right corner) - click "Last 1 hour" (in the list in the middle) - click "Auto-refresh" (top-right corner) - click "5 seconds" (top-left of the list) - You should see a series of green bars (with one new green bar every minute) .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## Updating our services to use GELF - We will now inform our Swarm to add GELF logging to all our services - This is done with the `docker service update` command - The logging flags are the same as before .lab[ - Enable GELF logging for the `rng` service: ```bash docker service update dockercoins_rng \ --log-driver gelf --log-opt gelf-address=udp://127.0.0.1:12201 ``` ] After ~15 seconds, you should see the log messages in Kibana. .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## Viewing container logs - Go back to Kibana - Container logs should be showing up! - We can customize the web UI to be more readable .lab[ - In the left column, move the mouse over the following columns, and click the "Add" button that appears: - host - container_name - message ] .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## .warning[Don't update stateful services!] - What would have happened if we had updated the Redis service? - When a service changes, SwarmKit replaces existing container with new ones - This is fine for stateless services - But if you update a stateful service, its data will be lost in the process - If we updated our Redis service, all our DockerCoins would be lost .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- ## Important afterword **This is not a "production-grade" setup.** It is just an educational example. We did set up a single ElasticSearch instance and a single Logstash instance. In a production setup, you need an ElasticSearch cluster (both for capacity and availability reasons). You also need multiple Logstash instances. And if you want to withstand bursts of logs, you need some kind of message queue: Redis if you're cheap, Kafka if you want to make sure that you don't drop messages on the floor. Good luck. If you want to learn more about the GELF driver, have a look at [this blog post]( https://jpetazzo.github.io/2017/01/20/docker-logging-gelf/). .debug[[swarm/logging.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/logging.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/containers-by-the-water.jpg)] --- name: toc-metrics-collection class: title Metrics collection .nav[ [Previous part](#toc-setting-up-elk-to-store-container-logs) | [Back to table of contents](#toc-part-4) | [Next part](#toc-dealing-with-stateful-services) ] .debug[(automatically generated title slide)] --- # Metrics collection - We want to gather metrics in a central place - We will gather node metrics and container metrics - We want a nice interface to view them (graphs) .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- ## Node metrics - CPU, RAM, disk usage on the whole node - Total number of processes running, and their states - Number of open files, sockets, and their states - I/O activity (disk, network), per operation or volume - Physical/hardware (when applicable): temperature, fan speed ... - ... and much more! .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- ## Container metrics - Similar to node metrics, but not totally identical - RAM breakdown will be different - active vs inactive memory - some memory is *shared* between containers, and accounted specially - I/O activity is also harder to track - async writes can cause deferred "charges" - some page-ins are also shared between containers For details about container metrics, see:
http://jpetazzo.github.io/2013/10/08/docker-containers-metrics/ .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- ## Application metrics - Arbitrary metrics related to your application and business - System performance: request latency, error rate ... - Volume information: number of rows in database, message queue size ... - Business data: inventory, items sold, revenue ... .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap, prom ## Tools We will build *two* different metrics pipelines: - One based on Intel Snap, - Another based on Prometheus. If you're using Play-With-Docker, skip the commands relevant to Intel Snap (we rely on a SSH server to deploy, and PWD doesn't have that yet). .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## First metrics pipeline We will use three open source Go projects for our first metrics pipeline: - Intel Snap Collects, processes, and publishes metrics - InfluxDB Stores metrics - Grafana Displays metrics visually .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Snap - [github.com/intelsdi-x/snap](https://github.com/intelsdi-x/snap) - Can collect, process, and publish metric data - Doesn’t store metrics - Works as a daemon (snapd) controlled by a CLI (snapctl) - Offloads collecting, processing, and publishing to plugins - Does nothing out of the box; configuration required! - Docs: https://github.com/intelsdi-x/snap/blob/master/docs/ .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## InfluxDB - Snap doesn't store metrics data - InfluxDB is specifically designed for time-series data - CRud vs. CRUD (you rarely if ever update/delete data) - orthogonal read and write patterns - storage format optimization is key (for disk usage and performance) - Snap has a plugin allowing to *publish* to InfluxDB .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Grafana - Snap cannot show graphs - InfluxDB cannot show graphs - Grafana will take care of that - Grafana can read data from InfluxDB and display it as graphs .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Getting and setting up Snap - We will install Snap directly on the nodes - Release tarballs are available from GitHub - We will use a *global service*
(started on all nodes, including nodes added later) - This service will download and unpack Snap in /opt and /usr/local - /opt and /usr/local will be bind-mounted from the host - This service will effectively install Snap on the hosts .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## The Snap installer service - This will get Snap on all nodes .lab[ ```bash docker service create --restart-condition=none --mode global \ --mount type=bind,source=/usr/local/bin,target=/usr/local/bin \ --mount type=bind,source=/opt,target=/opt centos sh -c ' SNAPVER=v0.16.1-beta RELEASEURL=https://github.com/intelsdi-x/snap/releases/download/$SNAPVER curl -sSL $RELEASEURL/snap-$SNAPVER-linux-amd64.tar.gz | tar -C /opt -zxf- curl -sSL $RELEASEURL/snap-plugins-$SNAPVER-linux-amd64.tar.gz | tar -C /opt -zxf- ln -s snap-$SNAPVER /opt/snap for BIN in snapd snapctl; do ln -s /opt/snap/bin/$BIN /usr/local/bin/$BIN; done ' # If you copy-paste that block, do not forget that final quote ☺ ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## First contact with `snapd` - The core of Snap is `snapd`, the Snap daemon - Application made up of a REST API, control module, and scheduler module .lab[ - Start `snapd` with plugin trust disabled and log level set to debug: ```bash snapd -t 0 -l 1 ``` ] - More resources: https://github.com/intelsdi-x/snap/blob/master/docs/SNAPD.md https://github.com/intelsdi-x/snap/blob/master/docs/SNAPD_CONFIGURATION.md .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Using `snapctl` to interact with `snapd` - Let's load a *collector* and a *publisher* plugins .lab[ - Open a new terminal - Load the psutil collector plugin: ```bash snapctl plugin load /opt/snap/plugin/snap-plugin-collector-psutil ``` - Load the file publisher plugin: ```bash snapctl plugin load /opt/snap/plugin/snap-plugin-publisher-mock-file ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Checking what we've done - Good to know: Docker CLI uses `ls`, Snap CLI uses `list` .lab[ - See your loaded plugins: ```bash snapctl plugin list ``` - See the metrics you can collect: ```bash snapctl metric list ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Actually collecting metrics: introducing *tasks* - To start collecting/processing/publishing metric data, you need to create a *task* - A *task* indicates: - *what* to collect (which metrics) - *when* to collect it (e.g. how often) - *how* to process it (e.g. use it directly, or compute moving averages) - *where* to publish it - Tasks can be defined with manifests written in JSON or YAML - Some plugins, such as the Docker collector, allow for wildcards (\*) in the metrics "path"
(see snap/docker-influxdb.json) - More resources: https://github.com/intelsdi-x/snap/blob/master/docs/TASKS.md .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Our first task manifest ```yaml version: 1 schedule: type: "simple" # collect on a set interval interval: "1s" # of every 1s max-failures: 10 workflow: collect: # first collect metrics: # metrics to collect /intel/psutil/load/load1: {} config: # there is no configuration publish: # after collecting, publish - plugin_name: "file" # use the file publisher config: file: "/tmp/snap-psutil-file.log" # write to this file ``` .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Creating our first task - The task manifest shown on the previous slide is stored in `snap/psutil-file.yml`. .lab[ - Create a task using the manifest: ```bash cd ~/container.training/snap snapctl task create -t psutil-file.yml ``` ] The output should look like the following: ``` Using task manifest to create task Task created ID: 240435e8-a250-4782-80d0-6fff541facba Name: Task-240435e8-a250-4782-80d0-6fff541facba State: Running ``` .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Checking existing tasks .lab[ - This will confirm that our task is running correctly, and remind us of its task ID ```bash snapctl task list ``` ] The output should look like the following: ``` ID NAME STATE HIT MISS FAIL CREATED 24043...acba Task-24043...acba Running 4 0 0 2:34PM 8-13-2016 ``` .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Viewing our task dollars at work - The task is using a very simple publisher, `mock-file` - That publisher just writes text lines in a file (one line per data point) .lab[ - Check that the data is flowing indeed: ```bash tail -f /tmp/snap-psutil-file.log ``` ] To exit, hit `^C` .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Debugging tasks - When a task is not directly writing to a local file, use `snapctl task watch` - `snapctl task watch` will stream the metrics you are collecting to STDOUT .lab[ ```bash snapctl task watch
``` ] To exit, hit `^C` .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Stopping snap - Our Snap deployment has a few flaws: - snapd was started manually - it is running on a single node - the configuration is purely local -- class: snap - We want to change that! -- class: snap - But first, go back to the terminal where `snapd` is running, and hit `^C` - All tasks will be stopped; all plugins will be unloaded; Snap will exit .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Snap Tribe Mode - Tribe is Snap's clustering mechanism - When tribe mode is enabled, nodes can join *agreements* - When a node in an *agreement* does something (e.g. load a plugin or run a task),
other nodes of that agreement do the same thing - We will use it to load the Docker collector and InfluxDB publisher on all nodes,
and run a task to use them - Without tribe mode, we would have to load plugins and run tasks manually on every node - More resources: https://github.com/intelsdi-x/snap/blob/master/docs/TRIBE.md .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Running Snap itself on every node - Snap runs in the foreground, so you need to use `&` or start it in tmux .lab[ - Run the following command *on every node*: ```bash snapd -t 0 -l 1 --tribe --tribe-seed node1:6000 ``` ] If you're *not* using Play-With-Docker, there is another way to start Snap! .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Starting a daemon through SSH .warning[Hackety hack ahead!] - We will create a *global service* - That global service will install a SSH client - With that SSH client, the service will connect back to its local node
(i.e. "break out" of the container, using the SSH key that we provide) - Once logged on the node, the service starts snapd with Tribe Mode enabled .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Running Snap itself on every node - I might go to hell for showing you this, but here it goes ... .lab[ - Start Snap all over the place: ```bash docker service create --name snapd --mode global \ --mount type=bind,source=$HOME/.ssh/id_rsa,target=/sshkey \ alpine sh -c " apk add --no-cache openssh-client && ssh -o StrictHostKeyChecking=no -i /sshkey docker@172.17.0.1 \ sudo snapd -t 0 -l 1 --tribe --tribe-seed node1:6000 " # If you copy-paste that block, don't forget that final quote :-) ``` ] Remember: this *does not work* with Play-With-Docker (which doesn't have SSH). .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Viewing the members of our tribe - If everything went fine, Snap is now running in tribe mode .lab[ - View the members of our tribe: ```bash snapctl member list ``` ] This should show the 5 nodes with their hostnames. .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Create an agreement - We can now create an *agreement* for our plugins and tasks .lab[ - Create an agreement; make sure to use the same name all along: ```bash snapctl agreement create docker-influxdb ``` ] The output should look like the following: ``` Name Number of Members plugins tasks docker-influxdb 0 0 0 ``` .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Instruct all nodes to join the agreement - We don't need another fancy global service! - We can join nodes from any existing node of the cluster .lab[ - Add all nodes to the agreement: ```bash snapctl member list | tail -n +2 | xargs -n1 snapctl agreement join docker-influxdb ``` ] The last bit of output should look like the following: ``` Name Number of Members plugins tasks docker-influxdb 5 0 0 ``` .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Start a container on every node - The Docker plugin requires at least one container to be started - Normally, at this point, you will have at least one container on each node - But just in case you did things differently, let's create a dummy global service .lab[ - Create an alpine container on the whole cluster: ```bash docker service create --name ping --mode global alpine ping 8.8.8.8 ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Running InfluxDB - We will create a service for InfluxDB - We will use the official image - InfluxDB uses multiple ports: - 8086 (HTTP API; we need this) - 8083 (admin interface; we need this) - 8088 (cluster communication; not needed here) - more ports for other protocols (graphite, collectd...) - We will just publish the first two .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Creating the InfluxDB service .lab[ - Start an InfluxDB service, publishing ports 8083 and 8086: ```bash docker service create --name influxdb \ --publish 8083:8083 \ --publish 8086:8086 \ influxdb:0.13 ``` ] Note: this will allow any node to publish metrics data to `localhost:8086`, and it will allow us to access the admin interface by connecting to any node on port 8083. .warning[Make sure to use InfluxDB 0.13; a few things changed in 1.0 (like, the name of the default retention policy is now "autogen") and this breaks a few things.] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Setting up InfluxDB - We need to create the "snap" database .lab[ - Open port 8083 with your browser - Enter the following query in the query box: ``` CREATE DATABASE "snap" ``` - In the top-right corner, select "Database: snap" ] Note: the InfluxDB query language *looks like* SQL but it's not. ??? ## Setting a retention policy - When graduating to 1.0, InfluxDB changed the name of the default policy - It used to be "default" and it is now "autogen" - Snap still uses "default" and this results in errors .lab[ - Create a "default" retention policy by entering the following query in the box: ``` CREATE RETENTION POLICY "default" ON "snap" DURATION 1w REPLICATION 1 ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Load Docker collector and InfluxDB publisher - We will load plugins on the local node - Since our local node is a member of the agreement, all other nodes in the agreement will also load these plugins .lab[ - Load Docker collector: ```bash snapctl plugin load /opt/snap/plugin/snap-plugin-collector-docker ``` - Load InfluxDB publisher: ```bash snapctl plugin load /opt/snap/plugin/snap-plugin-publisher-influxdb ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Start a simple collection task - Again, we will create a task on the local node - The task will be replicated on other nodes members of the same agreement .lab[ - Load a task manifest file collecting a couple of metrics on all containers,
and sending them to InfluxDB: ```bash cd ~/container.training/snap snapctl task create -t docker-influxdb.json ``` ] Note: the task description sends metrics to the InfluxDB API endpoint located at 127.0.0.1:8086. Since the InfluxDB container is published on port 8086, 127.0.0.1:8086 always routes traffic to the InfluxDB container. .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## If things go wrong... Note: if a task runs into a problem (e.g. it's trying to publish to a metrics database, but the database is unreachable), the task will be stopped. You will have to restart it manually by running: ```bash snapctl task enable
snapctl task start
``` This must be done *per node*. Alternatively, you can delete+re-create the task (it will delete+re-create on all nodes). .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Check that metric data shows up in InfluxDB - Let's check existing data with a few manual queries in the InfluxDB admin interface .lab[ - List "measurements": ``` SHOW MEASUREMENTS ``` (This should show two generic entries corresponding to the two collected metrics.) - View time series data for one of the metrics: ``` SELECT * FROM "intel/docker/stats/cgroups/cpu_stats/cpu_usage/total_usage" ``` (This should show a list of data points with **time**, **docker_id**, **source**, and **value**.) ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Deploy Grafana - We will use an almost-official image, `grafana/grafana` - We will publish Grafana's web interface on its default port (3000) .lab[ - Create the Grafana service: ```bash docker service create --name grafana --publish 3000:3000 grafana/grafana:3.1.1 ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Set up Grafana .lab[ - Open port 3000 with your browser - Identify with "admin" as the username and password - Click on the Grafana logo (the orange spiral in the top left corner) - Click on "Data Sources" - Click on "Add data source" (green button on the right) ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Add InfluxDB as a data source for Grafana .small[ Fill the form exactly as follows: - Name = "snap" - Type = "InfluxDB" In HTTP settings, fill as follows: - Url = "http://(IP.address.of.any.node):8086" - Access = "direct" - Leave HTTP Auth untouched In InfluxDB details, fill as follows: - Database = "snap" - Leave user and password blank Finally, click on "add", you should see a green message saying "Success - Data source is working". If you see an orange box (sometimes without a message), it means that you got something wrong. Triple check everything again. ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ![Screenshot showing how to fill the form](images/grafana-add-source.png) .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Create a dashboard in Grafana .lab[ - Click on the Grafana logo again (the orange spiral in the top left corner) - Hover over "Dashboards" - Click "+ New" - Click on the little green rectangle that appeared in the top left - Hover over "Add Panel" - Click on "Graph" ] At this point, you should see a sample graph showing up. .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ## Setting up a graph in Grafana .lab[ - Panel data source: select snap - Click on the SELECT metrics query to expand it - Click on "select measurement" and pick CPU usage - Click on the "+" right next to "WHERE" - Select "docker_id" - Select the ID of a container of your choice (e.g. the one running InfluxDB) - Click on the "+" on the right of the "SELECT" line - Add "derivative" - In the "derivative" option, select "1s" - In the top right corner, click on the clock, and pick "last 5 minutes" ] Congratulations, you are viewing the CPU usage of a single container! .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap ![Screenshot showing the end result](images/grafana-add-graph.png) .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap, prom ## Before moving on ... - Leave that tab open! - We are going to set up *another* metrics system - ... And then compare both graphs side by side .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: snap, prom ## Prometheus vs. Snap - Prometheus is another metrics collection system - Snap *pushes* metrics; Prometheus *pulls* them .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## Prometheus components - The *Prometheus server* pulls, stores, and displays metrics - Its configuration defines a list of *exporter* endpoints
(that list can be dynamic, using e.g. Consul, DNS, Etcd...) - The exporters expose metrics over HTTP using a simple line-oriented format (An optimized format using protobuf is also possible) .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## It's all about the `/metrics` - This is what the *node exporter* looks like: http://demo.robustperception.io:9100/metrics - Prometheus itself exposes its own internal metrics, too: http://demo.robustperception.io:9090/metrics - A *Prometheus server* will *scrape* URLs like these (It can also use protobuf to avoid the overhead of parsing line-oriented formats!) .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-manual ## Collecting metrics with Prometheus on Swarm - We will run two *global services* (i.e. scheduled on all our nodes): - the Prometheus *node exporter* to get node metrics - Google's cAdvisor to get container metrics - We will run a Prometheus server to scrape these exporters - The Prometheus server will be configured to use DNS service discovery - We will use `tasks.
` for service discovery - All these services will be placed on a private internal network .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-manual ## Creating an overlay network for Prometheus - This is the easiest step ☺ .lab[ - Create an overlay network: ```bash docker network create --driver overlay prom ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-manual ## Running the node exporter - The node exporter *should* run directly on the hosts - However, it can run from a container, if configured properly
(it needs to access the host's filesystems, in particular /proc and /sys) .lab[ - Start the node exporter: ```bash docker service create --name node --mode global --network prom \ --mount type=bind,source=/proc,target=/host/proc \ --mount type=bind,source=/sys,target=/host/sys \ --mount type=bind,source=/,target=/rootfs \ prom/node-exporter \ --path.procfs /host/proc \ --path.sysfs /host/sys \ --collector.filesystem.ignored-mount-points "^/(sys|proc|dev|host|etc)($|/)" ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-manual ## Running cAdvisor - Likewise, cAdvisor *should* run directly on the hosts - But it can run in containers, if configured properly .lab[ - Start the cAdvisor collector: ```bash docker service create --name cadvisor --network prom --mode global \ --mount type=bind,source=/,target=/rootfs \ --mount type=bind,source=/var/run,target=/var/run \ --mount type=bind,source=/sys,target=/sys \ --mount type=bind,source=/var/lib/docker,target=/var/lib/docker \ google/cadvisor:latest ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-manual ## Prometheus server configuration This will be our configuration file for Prometheus: .small[ ```yaml global: scrape_interval: 10s scrape_configs: - job_name: 'prometheus' static_configs: - targets: ['localhost:9090'] - job_name: 'node' dns_sd_configs: - names: ['tasks.node'] type: 'A' port: 9100 - job_name: 'cadvisor' dns_sd_configs: - names: ['tasks.cadvisor'] type: 'A' port: 8080 ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-manual ## Passing the configuration to Prometheus - The easiest solution is to create a custom image bundling this configuration - We will use a very simple Dockerfile: ```dockerfile FROM prom/prometheus:v1.4.1 COPY prometheus.yml /etc/prometheus/prometheus.yml ``` (The configuration file, and the Dockerfile, are in the `prom` subdirectory) - We will build this image, and push it to our local registry - Then we will create a service using this image Note: it is also possible to use a `config` to inject that configuration file without having to create this ad-hoc image. .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-manual ## Building our custom Prometheus image - We will use the local registry started previously on 127.0.0.1:5000 .lab[ - Build the image using the provided Dockerfile: ```bash docker build -t 127.0.0.1:5000/prometheus ~/container.training/prom ``` - Push the image to our local registry: ```bash docker push 127.0.0.1:5000/prometheus ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-manual ## Running our custom Prometheus image - That's the only service that needs to be published (If we want to access Prometheus from outside!) .lab[ - Start the Prometheus server: ```bash docker service create --network prom --name prom \ --publish 9090:9090 127.0.0.1:5000/prometheus ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-auto ## Deploying Prometheus on our cluster - We will use a stack definition (once again) .lab[ - Make sure we are in the stacks directory: ```bash cd ~/container.training/stacks ``` - Build, ship, and run the Prometheus stack: ```bash docker-compose -f prometheus.yml build docker-compose -f prometheus.yml push docker stack deploy -c prometheus.yml prometheus ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## Checking our Prometheus server - First, let's make sure that Prometheus is correctly scraping all metrics .lab[ - Open port 9090 with your browser - Click on "status", then "targets" ] You should see 7 endpoints (3 cadvisor, 3 node, 1 prometheus). Their state should be "UP". .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-auto, config ## Injecting a configuration file (New in Docker Engine 17.06) - We are creating a custom image *just to inject a configuration* - Instead, we could use the base Prometheus image + a `config` - A `config` is a blob (usually, a configuration file) that: - is created and managed through the Docker API (and CLI) - gets persisted into the Raft log (i.e. safely) - can be associated to a service
(this injects the blob as a plain file in the service's containers) .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-auto, config ## Differences between `config` and `secret` The two are very similar, but ... - `configs`: - can be injected to any filesystem location - can be viewed and extracted using the Docker API or CLI - `secrets`: - can only be injected into `/run/secrets` - are never stored in clear text on disk - cannot be viewed or extracted with the Docker API or CLI .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-auto, config ## Deploying Prometheus with a `config` The following Compose file (`prometheus+config.yml`) achieves the same result, but by using a `config` instead of baking the configuration into the image. .small[ ```yaml version: "3.3" services: prometheus: image: prom/prometheus:v1.4.1 ports: - "9090:9090" configs: - source: prometheus target: /etc/prometheus/prometheus.yml ... configs: prometheus: file: ../prom/prometheus.yml ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-auto, config ## Specifying a `config` in a Compose file - In each service, an optional `configs` section can list as many configs as you want - Each config can specify: - an optional `target` (path to inject the configuration; by default: root of the container) - ownership and permissions (by default, the file will be owned by UID 0, i.e. `root`) - These configs reference top-level `configs` elements - The top-level configs can be declared as: - *external*, meaning that it is supposed to be created before you deploy the stack - referencing a file, whose content is used to initialize the config .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-auto, config ## Re-deploying Prometheus with a config - We will update the existing stack using `prometheus+config.yml` .lab[ - Redeploy the `prometheus` stack: ```bash docker stack deploy -c prometheus+config.yml prometheus ``` - Check that Prometheus still works as intended (By connecting to any node of the cluster, on port 9090) ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-auto, config ## Accessing the config object from the CLI - Config objects can be viewed from the Docker CLI (or API) .lab[ - List existing config objects: ```bash docker config ls ``` - View details about our config object: ```bash docker config inspect prometheus_prometheus ``` ] Note: the content of the config blob is shown with BASE64 encoding.
(It doesn't have to be text; it could be an image or any kind of binary content!) .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom-auto, config ## Extracting a config blob - Let's retrieve that Prometheus configuration! .lab[ - Extract the BASE64 payload with `jq`: ```bash docker config inspect prometheus_prometheus | jq -r .[0].Spec.Data ``` - Decode it with `base64 -d`: ```bash docker config inspect prometheus_prometheus | jq -r .[0].Spec.Data | base64 -d ``` ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## Displaying metrics directly from Prometheus - This is easy ... if you are familiar with PromQL .lab[ - Click on "Graph", and in "expression", paste the following: ``` sum by (container_label_com_docker_swarm_node_id) ( irate( container_cpu_usage_seconds_total{ container_label_com_docker_swarm_service_name="dockercoins_worker" }[1m] ) ) ``` - Click on the blue "Execute" button and on the "Graph" tab just below ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## Building the query from scratch - We are going to build the same query from scratch - This doesn't intend to be a detailed PromQL course - This is merely so that you (I) can pretend to know how the previous query works
so that your coworkers (you) can be suitably impressed (or not) (Or, so that we can build other queries if necessary, or adapt if cAdvisor, Prometheus, or anything else changes and requires editing the query!) .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## Displaying a raw metric for *all* containers - Click on the "Graph" tab on top *This takes us to a blank dashboard* - Click on the "Insert metric at cursor" drop down, and select `container_cpu_usage_seconds_total` *This puts the metric name in the query box* - Click on "Execute" *This fills a table of measurements below* - Click on "Graph" (next to "Console") *This replaces the table of measurements with a series of graphs (after a few seconds)* .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## Selecting metrics for a specific service - Hover over the lines in the graph (Look for the ones that have labels like `container_label_com_docker_...`) - Edit the query, adding a condition between curly braces: .small[`container_cpu_usage_seconds_total{container_label_com_docker_swarm_service_name="dockercoins_worker"}`] - Click on "Execute" *Now we should see one line per CPU per container* - If you want to select by container ID, you can use a regex match: `id=~"/docker/c4bf.*"` - You can also specify multiple conditions by separating them with commas .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## Turn counters into rates - What we see is the total amount of CPU used (in seconds) - We want to see a *rate* (CPU time used / real time) - To get a moving average over 1 minute periods, enclose the current expression within: ``` rate ( ... { ... } [1m] ) ``` *This should turn our steadily-increasing CPU counter into a wavy graph* - To get an instantaneous rate, use `irate` instead of `rate` (The time window is then used to limit how far behind to look for data if data points are missing in case of scrape failure; see [here](https://www.robustperception.io/irate-graphs-are-better-graphs/) for more details!) *This should show spikes that were previously invisible because they were smoothed out* .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## Aggregate multiple data series - We have one graph per CPU per container; we want to sum them - Enclose the whole expression within: ``` sum ( ... ) ``` *We now see a single graph* .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom ## Collapse dimensions - If we have multiple containers we can also collapse just the CPU dimension: ``` sum without (cpu) ( ... ) ``` *This shows the same graph, but preserves the other labels* - Congratulations, you wrote your first PromQL expression from scratch! (I'd like to thank [Johannes Ziemke](https://twitter.com/discordianfish) and [Julius Volz](https://twitter.com/juliusvolz) for their help with Prometheus!) .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom, snap ## Comparing Snap and Prometheus data - If you haven't set up Snap, InfluxDB, and Grafana, skip this section - If you have closed the Grafana tab, you might have to re-set up a new dashboard (Unless you saved it before navigating it away) - To re-do the setup, just follow again the instructions from the previous chapter .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom, snap ## Add Prometheus as a data source in Grafana .lab[ - In a new tab, connect to Grafana (port 3000) - Click on the Grafana logo (the orange spiral in the top-left corner) - Click on "Data Sources" - Click on the green "Add data source" button ] We see the same input form that we filled earlier to connect to InfluxDB. .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom, snap ## Connecting to Prometheus from Grafana .lab[ - Enter "prom" in the name field - Select "Prometheus" as the source type - Enter http://(IP.address.of.any.node):9090 in the Url field - Select "direct" as the access method - Click on "Save and test" ] Again, we should see a green box telling us "Data source is working." Otherwise, double-check every field and try again! .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom, snap ## Adding the Prometheus data to our dashboard .lab[ - Go back to the the tab where we had our first Grafana dashboard - Click on the blue "Add row" button in the lower right corner - Click on the green tab on the left; select "Add panel" and "Graph" ] This takes us to the graph editor that we used earlier. .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom, snap ## Querying Prometheus data from Grafana The editor is a bit less friendly than the one we used for InfluxDB. .lab[ - Select "prom" as Panel data source - Paste the query in the query field: ``` sum without (cpu, id) ( irate ( container_cpu_usage_seconds_total{ container_label_com_docker_swarm_service_name="influxdb"}[1m] ) ) ``` - Click outside of the query field to confirm - Close the row editor by clicking the "X" in the top right area ] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: prom, snap ## Interpreting results - The two graphs *should* be similar - Protip: align the time references! .lab[ - Click on the clock in the top right corner - Select "last 30 minutes" - Click on "Zoom out" - Now press the right arrow key (hold it down and watch the CPU usage increase!) ] *Adjusting units is left as an exercise for the reader.* .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- ## More resources on container metrics - [Prometheus, a Whirlwind Tour](https://speakerdeck.com/copyconstructor/prometheus-a-whirlwind-tour), an original overview of Prometheus - [Docker Swarm & Container Overview](https://grafana.net/dashboards/609), a custom dashboard for Grafana - [Gathering Container Metrics](http://jpetazzo.github.io/2013/10/08/docker-containers-metrics/), a blog post about cgroups - [The Prometheus Time Series Database](https://www.youtube.com/watch?v=HbnGSNEjhUc), a talk explaining why custom data storage is necessary for metrics .blackbelt[DC17US: Monitoring, the Prometheus Way ([video](https://www.youtube.com/watch?v=PDxcEzu62jk&list=PLkA60AVN3hh-biQ6SCtBJ-WVTyBmmYho8&index=5))] .blackbelt[DC17EU: Prometheus 2.0 Storage Engine ([video](https://dockercon.docker.com/watch/NNZ8GXHGomouwSXtXnxb8P))] .debug[[swarm/metrics.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/metrics.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/distillery-containers.jpg)] --- name: toc-dealing-with-stateful-services class: title Dealing with stateful services .nav[ [Previous part](#toc-metrics-collection) | [Back to table of contents](#toc-part-4) | [Next part](#toc-controlling-docker-from-a-container) ] .debug[(automatically generated title slide)] --- # Dealing with stateful services - First of all, you need to make sure that the data files are on a *volume* - Volumes are host directories that are mounted to the container's filesystem - These host directories can be backed by the ordinary, plain host filesystem ... - ... Or by distributed/networked filesystems - In the latter scenario, in case of node failure, the data is safe elsewhere ... - ... And the container can be restarted on another node without data loss .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Building a stateful service experiment - We will use Redis for this example - We will expose it on port 10000 to access it easily .lab[ - Start the Redis service: ```bash docker service create --name stateful -p 10000:6379 redis ``` - Check that we can connect to it: ```bash docker run --net host --rm redis redis-cli -p 10000 info server ``` ] .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Accessing our Redis service easily - Typing that whole command is going to be tedious .lab[ - Define a shell alias to make our lives easier: ```bash alias redis='docker run --net host --rm redis redis-cli -p 10000' ``` - Try it: ```bash redis info server ``` ] .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Basic Redis commands .lab[ - Check that the `foo` key doesn't exist: ```bash redis get foo ``` - Set it to `bar`: ```bash redis set foo bar ``` - Check that it exists now: ```bash redis get foo ``` ] .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Local volumes vs. global volumes - Global volumes exist in a single namespace - A global volume can be mounted on any node
.small[(bar some restrictions specific to the volume driver in use; e.g. using an EBS-backed volume on a GCE/EC2 mixed cluster)] - Attaching a global volume to a container allows to start the container anywhere
(and retain its data wherever you start it!) - Global volumes require extra *plugins* (Flocker, Portworx...) - Docker doesn't come with a default global volume driver at this point - Therefore, we will fall back on *local volumes* .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Local volumes - We will use the default volume driver, `local` - As the name implies, the `local` volume driver manages *local* volumes - Since local volumes are (duh!) *local*, we need to pin our container to a specific host - We will do that with a *constraint* .lab[ - Add a placement constraint to our service: ```bash docker service update stateful --constraint-add node.hostname==$HOSTNAME ``` ] .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Where is our data? - If we look for our `foo` key, it's gone! .lab[ - Check the `foo` key: ```bash redis get foo ``` - Adding a constraint caused the service to be redeployed: ```bash docker service ps stateful ``` ] Note: even if the constraint ends up being a no-op (i.e. not moving the service), the service gets redeployed. This ensures consistent behavior. .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Setting the key again - Since our database was wiped out, let's populate it again .lab[ - Set `foo` again: ```bash redis set foo bar ``` - Check that it's there: ```bash redis get foo ``` ] .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Updating a service recreates its containers - Let's try to make a trivial update to the service and see what happens .lab[ - Set a memory limit to our Redis service: ```bash docker service update stateful --limit-memory 100M ``` - Try to get the `foo` key one more time: ```bash redis get foo ``` ] The key is blank again! .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Service volumes are ephemeral by default - Let's highlight what's going on with volumes! .lab[ - Check the current list of volumes: ```bash docker volume ls ``` - Carry a minor update to our Redis service: ```bash docker service update stateful --limit-memory 200M ``` ] Again: all changes trigger the creation of a new task, and therefore a replacement of the existing container; even when it is not strictly technically necessary. .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## The data is gone again - What happened to our data? .lab[ - The list of volumes is slightly different: ```bash docker volume ls ``` ] (You should see one extra volume.) .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Assigning a persistent volume to the container - Let's add an explicit volume mount to our service, referencing a named volume .lab[ - Update the service with a volume mount: ```bash docker service update stateful \ --mount-add type=volume,source=foobarstore,target=/data ``` - Check the new volume list: ```bash docker volume ls ``` ] Note: the `local` volume driver automatically creates volumes. .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Checking that data is now persisted correctly .lab[ - Store something in the `foo` key: ```bash redis set foo barbar ``` - Update the service with yet another trivial change: ```bash docker service update stateful --limit-memory 300M ``` - Check that `foo` is still set: ```bash redis get foo ``` ] .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Recap - The service must commit its state to disk when being shutdown.red[*] (Shutdown = being sent a `TERM` signal) - The state must be written on files located on a volume - That volume must be specified to be persistent - If using a local volume, the service must also be pinned to a specific node (And losing that node means losing the data, unless there are other backups) .footnote[
.red[*]If you customize Redis configuration, make sure you persist data correctly!
It's easy to make that mistake — __Trust me!__] .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Cleaning up .lab[ - Remove the stateful service: ```bash docker service rm stateful ``` - Remove the associated volume: ```bash docker volume rm foobarstore ``` ] Note: we could keep the volume around if we wanted. .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- ## Should I run stateful services in containers? -- Depending whom you ask, they'll tell you: -- - certainly not, heathen! -- - we've been running a few thousands PostgreSQL instances in containers ...
for a few years now ... in production ... is that bad? -- - what's a container? -- Perhaps a better question would be: *"Should I run stateful services?"* -- - is it critical for my business? - is it my value-add? - or should I find somebody else to run them for me? .debug[[swarm/stateful.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/stateful.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/lots-of-containers.jpg)] --- name: toc-controlling-docker-from-a-container class: title Controlling Docker from a container .nav[ [Previous part](#toc-dealing-with-stateful-services) | [Back to table of contents](#toc-part-4) | [Next part](#toc-node-management) ] .debug[(automatically generated title slide)] --- # Controlling Docker from a container - In a local environment, just bind-mount the Docker control socket: ```bash docker run -ti -v /var/run/docker.sock:/var/run/docker.sock docker ``` - Otherwise, you have to: - set `DOCKER_HOST`, - set `DOCKER_TLS_VERIFY` and `DOCKER_CERT_PATH` (if you use TLS), - copy certificates to the container that will need API access. More resources on this topic: - [Do not use Docker-in-Docker for CI]( https://jpetazzo.github.io/2015/09/03/do-not-use-docker-in-docker-for-ci/) - [One container to rule them all]( https://jpetazzo.github.io/2016/04/03/one-container-to-rule-them-all/) .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Bind-mounting the Docker control socket - In Swarm mode, bind-mounting the control socket gives you access to the whole cluster - You can tell Docker to place a given service on a manager node, using constraints: ```bash docker service create \ --mount source=/var/run/docker.sock,type=bind,target=/var/run/docker.sock \ --name autoscaler --constraint node.role==manager ... ``` .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Constraints and global services (New in Docker Engine 1.13) - By default, global services run on *all* nodes ```bash docker service create --mode global ... ``` - You can specify constraints for global services - These services will run only on the node satisfying the constraints - For instance, this service will run on all manager nodes: ```bash docker service create --mode global --constraint node.role==manager ... ``` .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Constraints and dynamic scheduling (New in Docker Engine 1.13) - If constraints change, services are started/stopped accordingly (e.g., `--constraint node.role==manager` and nodes are promoted/demoted) - This is particularly useful with labels: ```bash docker node update node1 --label-add defcon=five docker service create --constraint node.labels.defcon==five ... docker node update node2 --label-add defcon=five docker node update node1 --label-rm defcon=five ``` .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Shortcomings of dynamic scheduling .warning[If a service becomes "unschedulable" (constraints can't be satisfied):] - It won't be scheduled automatically when constraints are satisfiable again - You will have to update the service; you can do a no-op update with: ```bash docker service update ... --force ``` .warning[Docker will silently ignore attempts to remove a non-existent label or constraint] - It won't warn you if you typo when removing a label or constraint! .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/plastic-containers.JPG)] --- name: toc-node-management class: title Node management .nav[ [Previous part](#toc-controlling-docker-from-a-container) | [Back to table of contents](#toc-part-4) | [Next part](#toc-links-and-resources) ] .debug[(automatically generated title slide)] --- # Node management - SwarmKit allows to change (almost?) everything on-the-fly - Nothing should require a global restart .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Node availability ```bash docker node update
--availability
``` - Active = schedule tasks on this node (default) - Pause = don't schedule new tasks on this node; existing tasks are not affected You can use it to troubleshoot a node without disrupting existing tasks It can also be used (in conjunction with labels) to reserve resources - Drain = don't schedule new tasks on this node; existing tasks are moved away This is just like crashing the node, but containers get a chance to shutdown cleanly .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Managers and workers - Nodes can be promoted to manager with `docker node promote` - Nodes can be demoted to worker with `docker node demote` - This can also be done with `docker node update
--role
` - Reminder: this has to be done from a manager node
(workers cannot promote themselves) .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Removing nodes - You can leave Swarm mode with `docker swarm leave` - Nodes are drained before being removed (i.e. all tasks are rescheduled somewhere else) - Managers cannot leave (they have to be demoted first) - After leaving, a node still shows up in `docker node ls` (in `Down` state) - When a node is `Down`, you can remove it with `docker node rm` (from a manager node) .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Join tokens and automation - If you have used Docker 1.12-RC: join tokens are now mandatory! - You cannot specify your own token (SwarmKit generates it) - If you need to change the token: `docker swarm join-token --rotate ...` - To automate cluster deployment: - have a seed node do `docker swarm init` if it's not already in Swarm mode - propagate the token to the other nodes (secure bucket, facter, ohai...) .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Viewing disk usage: `docker system df` (New in Docker Engine 1.13) - Shows disk usage for images, containers, and volumes - Breaks down between *active* and *reclaimable* categories .lab[ - Check how much disk space is used at the end of the workshop: ```bash docker system df ``` ] .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Cleaning up disk: `docker system prune` - Removes stopped containers - Removes dangling images (that don't have a tag associated anymore) - Removes orphaned volumes - Removes empty networks .lab[ - Try it: ```bash docker system prune -f ``` ] Note: `docker system prune -a` will also remove *unused* images. .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Events - You can get a real-time stream of events with `docker events` - This will report *local events* and *cluster events* - Local events =
all activity related to containers, images, plugins, volumes, networks, *on this node* - Cluster events =
Swarm Mode activity related to services, nodes, secrets, configs, *on the whole cluster* - `docker events` doesn't report *local events happening on other nodes* - Events can be filtered (by type, target, labels...) - Events can be formatted with Go's `text/template` or in JSON .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- ## Getting *all the events* - There is no built-in to get a stream of *all the events* on *all the nodes* - This can be achieved with (for instance) the four following services working together: - a Redis container (used as a stateless, fan-in message queue) - a global service bind-mounting the Docker socket, pushing local events to the queue - a similar singleton service to push global events to the queue - a queue consumer fetching events and processing them as you please I'm not saying that you should implement it with Shell scripts, but you totally could. .small[ (It might or might not be one of the initiating rites of the [House of Bash](https://twitter.com/carmatrocity/status/676559402787282944)) ] For more information about event filters and types, check [the documentation](https://docs.docker.com/engine/reference/commandline/events/). .debug[[swarm/extratips.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/extratips.md)] --- class: title, self-paced Thank you! .debug[[shared/thankyou.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/thankyou.md)] --- class: title, in-person That's all, folks!
Questions? ![end](images/end.jpg) .debug[[shared/thankyou.md](https://github.com/jpetazzo/container.training/tree/main/slides/shared/thankyou.md)] --- class: pic .interstitial[![Image separating from the next part](https://prettypictures.container.training/containers/train-of-containers-1.jpg)] --- name: toc-links-and-resources class: title Links and resources .nav[ [Previous part](#toc-node-management) | [Back to table of contents](#toc-part-4) | [Next part](#toc-) ] .debug[(automatically generated title slide)] --- # Links and resources - [Docker Community Slack](https://community.docker.com/registrations/groups/4316) - [Docker Community Forums](https://forums.docker.com/) - [Docker Hub](https://hub.docker.com) - [Docker Blog](https://blog.docker.com/) - [Docker documentation](https://docs.docker.com/) - [Docker on StackOverflow](https://stackoverflow.com/questions/tagged/docker) - [Docker on Twitter](https://twitter.com/docker) - [Play With Docker Hands-On Labs](https://training.play-with-docker.com/) .footnote[These slides (and future updates) are on → https://container.training/] .debug[[swarm/links.md](https://github.com/jpetazzo/container.training/tree/main/slides/swarm/links.md)]