In this tutorial I want to show you how to build a 3-node Kubernetes cluster with Rancher's k3s project and k3sup (ketchup), which uses ssh
to make the whole process quick and painless.
Introduction
At time of writing, k3s is one of the most popular installable Kubernetes distributions on the CNCF Landscape
I am a fan of managed Kubernetes services, but I am also a fan of choice and there is no managed Kubernetes service for k3s on DigitalOcean at this time.
Now you may ask me why we need k3s, if DigitalOcean has its own Kubernetes service in GA. k3s takes a unique approach of stripping-down Kubernetes to its bare essentials and makes it perfect for running in edge locations. Given that less resources are needed, the spend may be lower, and a LoadBalancer (10-15 USD) is not required because of the way k3s exposes traffic through HostPorts.
The k3sup community has been building up a series of tutorials for various developer clouds and VM platforms. DigitalOcean is a great place to test both k3s and those tutorials given its fast launch time and simple user experience.
Still not convinced that saving 10-15 USD / month is worth it? Read on, or feel free to try k3sup on another cloud using one of the other community blog posts.
Provision the VMs
Set up three VMs using the DigitalOcean UI (control-panel).
I decided to get a balance of 2vCPU and 2GB RAM. You can pick whatever works for you.
Configure your SSH key, so that k3sup can use it later on to orchestrate the nodes:
I've also added private networking.
You can save a tag in this last step for easier management, or just use names that have a prefix:
Setup the DigitalOcean CLI
In my experience, it's easier to create your droplets using the UI and then to manage them through the CLI.
-
Install doctl
-
Create an API key in the UI with Read and Write access
-
Authenticate the CLI -
doctl auth init
List your VMs:
doctl compute droplet ls --format Name,PublicIPv4,Status
Name Public IPv4 Status
k3s-1 159.65.25.233 active
k3s-2 159.65.22.30 active
k3s-3 159.65.29.53 active
Install k3sup
Let's now install k3sup
(ketchup), which has one job, to bring k3s up.
curl -sLSf https://get.k3sup.dev | sh
sudo install -m k3sup /usr/local/bin/
If you're a Windows user, then checkout the Releases page on GitHub.
Create the cluster
Here's a conceptual diagram of what we will be doing:
k3sup
uses ssh
to run commands on the remote machines, which is a similar approach to automation tooling such as Ansible.
Create the server
In Kubernetes terminology, the server is often called the master. The master node is prevented from running workloads and then an additional command is needed to "untaint" the master. k3s is designed for the edge where you may only have a single server, it can run workloads out of the box.
export SERVER_IP=159.65.25.233
k3sup install --ip $SERVER_IP --user root --local-path=./kubeconfig
You can alter the --user
flag if needed.
This will save a file to the current working directory.
k3s is so fast to start up, that it may be ready for use after the command has completed.
- Test it out:
export KUBECONFIG=`pwd`/kubeconfig
kubectl get node -o wide
NAME STATUS ROLES AGE VERSION INTERNAL-IP EXTERNAL-IP OS-IMAGE KERNEL-VERSION CONTAINER-RUNTIME
k3s-1 Ready master 22s v1.14.6-k3s.1 159.65.25.233 <none> Ubuntu 18.04.3 LTS 4.15.0-58-generic containerd://1.2.7-k3s1
- Run a test pod:
This pod ships with curl
pre-installed.
kubectl run -t -i curl --image alexellis/curl:0.1.1 -- /bin/sh
~ $
~ $ curl 'https://api6.ipify.org?format=json' ; echo
{"ip":"159.65.25.233"}
-
Type
exit
-
Delete the pod:
kubectl delete deploy/curl
If you ever lose your kubeconfig file, then you can run k3sup install --skip-install
to simply copy it back to your computer again.
Extend the cluster
You could stop here and run your workloads all on a single node, but adding additional hosts will expand our available capacity.
k3s-2 159.65.22.30
k3s-3 159.65.29.53
Run the following:
export SERVER_IP=159.65.25.233
# k3s-2
export IP=159.65.22.30
k3sup join --ip $IP --server-ip $SERVER_IP --user root
#k3s-3
export IP=159.65.29.53
k3sup join --ip $IP --server-ip $SERVER_IP --user root
Check the cluster
You'll now find 3 nodes listed as "Ready"
kubectl get node
NAME STATUS ROLES AGE VERSION
k3s-1 Ready master 6m57s v1.14.6-k3s.1
k3s-2 Ready worker 55s v1.14.6-k3s.1
k3s-3 Ready worker 7s v1.14.6-k3s.1
You can find out what workloads are running on each host and how they are doing for memory/CPU/disk pressure by running:
kubectl describe node/k3s-1
Secure the cluster
Familiarise yourself with how to secure k3s for the public internet by reading the recommendations in the official documentation: Open Ports and Network Security
Wrapping up
We now have a cluster with 3 nodes and a collective capacity of 6vCPUs and 6GB RAM. You can deploy any helm chart to the cluster and start running your workloads there. K3s comes with an IngressController called Traefik, and from my experience it's easier to go with that than to try to install your favourite.
The k3sup
tool works with any VM, cloud or even bare-metal like Raspberry Pi. Try one of the community tutorials, or write one for your favourite VPS.
If you liked k3sup, and want to support my work, become a GitHub Sponsor or follow me on Twitter.
Take it further
In a follow-up post, I'll show you how to set up your own domain-name and how to add TLS to a HTTP API.
Here are some of my recommended charts and Kubernetes software:
Essentials:
- openfaas - a next-generation FaaS / PaaS making it easy to deploy applications without having to learn all of the ins and outs of Kubernetes
- cert-manager - protect your services with HTTPS/TLS
- postgres - "PostgreSQL: The world's most advanced open source database"
Nice to have / advanced:
- inlets - expose your local clusters and servers via the public IP addresses of your Kubernetes cluster, a bit like an Open Source version of ngrok
- weave scope - explore your network traffic and resources visually
- jenkins - open source CI/CD tooling for building software releases - also useful for creating Docker images
- longhorn - add PersistentVolume / Block storage support to k3s
version of ngrok.
You can install most of them using the Open Source Kubernetes marketplace, arkade.
For example:
arkade install jenkins
arkade install postgres
arkade install openfaas
You'll get instructions on how to use each app after the installation, and can get the information back at any time with arkade info APP
.