The user-facing API changes more rarely, usually as a result of a Kubernetes version upgrade. For details, read the user-facing release notes.
bin
: wrapper scripts that helps you run kubesprayconfig
: default config valuesconformance-tests
: ck8s conformance tests for bare metal machineskubespray
: git submodule of the kubespray repository
terraform (tested with 1.2.9)
Installs requirements using the ansible playbook get-requirements.yaml
ansible-playbook -e 'ansible_python_interpreter=/usr/bin/python3' --ask-become-pass --connection local --inventory 127.0.0.1, get-requirements.yaml
-
Init the kubespray config in your config path
export CK8S_CONFIG_PATH=~/.ck8s/my-environment ./bin/ck8s-kubespray init <wc|sc> <flavor> [<SOPS fingerprint>]
Arguments:
- The init command accepts
wc
(workload cluster) orsc
(service cluster) as first argument as to create separate folders for each cluster's configuration files. flavor
will determine some default values for a variety of config options. Supported options aredefault
,gcp
,aws
,vsphere
, andopenstack
.SOPS fingerprint
is the gpg fingerprint that will be used for SOPS encryption. You need to set this or the environment variableCK8S_PGP_FP
the first time SOPS is used in your specified config path.
- The init command accepts
-
Edit the
inventory.ini
(found in your config path) to match the VMs (IP addresses and other settings that might be needed for your setup) that should be part of the cluster. Or if you have one created by a terraform script inkubespray/contrib/terraform
you should use that one. -
Init and update the kubespray gitsubmodule:
git submodule init git submodule update
-
Run kubespray to set up the kubernetes cluster:
./bin/ck8s-kubespray apply <wc|sc> [<options>]
Any
options
added will be forwarded to ansible. -
Done. You should now have a working kubernetes cluster. You should also have an encrypted kubeconfig at
<CK8S_CONFIG_PATH>/.state/kube_config_<wc|sc>.yaml
that you can use to access the cluster.
Authorized SSH keys can be changed for a cluster using:
./bin/ck8s-kubespray apply-ssh <wc|sc> [<options>]
It will set the public SSH key(s) found in<CK8S_CONFIG_PATH>/<wc|sc>-config/group_vars/all/ck8s-ssh-keys.yaml
as authorized keys in your cluster (just add the keys you want to be authorized as elements in ck8s_ssh_pub_keys_list
).
Note that the authorized SSH keys for the cluster will be set to these keys exclusively, removing any keys that may already be authorized, so make sure the list includes every SSH key that should be authorized.
When running this command, the SSH keys are applied to each node in the cluster sequentially, in reverse inventory order (first the workers and then the masters). A connection test is performed after each node which has to succeed in order for the playbook to continue. If the connection test fails, you may have lost your SSH access to the node; to recover from this, you can set up an SSH connection before running the command and keep it active so that you can change the authorized keys manually.
You can reboot all nodes that wants to restart (usually to finish installing new packages) by running:
./bin/ck8s-kubespray reboot-nodes <wc|sc> [--extra-vars manual_prompt=true] [<options>]
If you set --extra-vars manual_prompt=true
then you get a manual prompt before each reboot so you can stop the playbook if you want.
Note that this playbook requires you to use ansible version >= 2.10.
You can remove a node from a ck8s cluster by running:
./bin/ck8s-kubespray remove-node <wc|sc> <node-name>[,<node-name-2>,...] [<options>]
- The script may fail with the message
error while evaluating conditional (kubelet_heartbeat.rc == 0): 'dict object' has no attribute 'rc'
- In such situations just rerun the script. It will skip the check for that node, so check that it is up and running manually afterwards.
- The script might fail with a timeout:
Timeout (12s) waiting for privilege escalation prompt
- Try running the script again with a longer ansible timeout:
export ANSIBLE_TIMEOUT=30
- Try running the script again with a longer ansible timeout:
With the following command you can run any ansible playbook available in kubespray:
./bin/ck8s-kubespray run-playbook <wc|sc> <playbook> [<options>]
Where playbook
is the filename of the playbook that you want to run, e.g. cluster.yml
if you want to create a cluster (making the command functionally the same as our ck8s-kubespray apply
command) or scale.yml
if you want to just add more nodes. Remember to check the kubespray documentation before running a playbook.
This will use the inventory, group-vars, and ssh key in your config path and therefore requires that you first run the init command. Any options
added will be forwarded to ansible.
We recommend that you use OIDC kubeconfigs instead of regular cluster-admin kubeconfigs. The default settings will create OIDC kubeconfigs for you when you run ./bin/ck8s-kubespray apply <wc|sc>
, but there are some variables you need to set. See the variables in: <wc|sc>-config/group_vars/k8s_cluster/ck8s-k8s-cluster.yaml
in your config path.
But if you need to use a regular cluster-admin kubeconfig in a break-glass situation, then you can ssh to one of the controleplane nodes and use the kubeconfig at /etc/kubernetes/admin.conf
. We recommend that you do not copy that kubeconfig to your local host, when dealing with production clusters.
For development you can skip OIDC and instead get a regular cluster admin kubeconfig locally by setting kubeconfig_localhost: true
. Note that you then must set create_oidc_kubeconfig: false
.
The kubeconfig and OIDC cluster admin RBAC are managed with the playbooks playbooks/kubeconfig.yml
and playbooks/cluster_admin_rbac.yml
. You can run them manually with:
./bin/ck8s-kubespray run-playbook <wc|sc> ../playbooks/kubeconfig.yml -b
./bin/ck8s-kubespray run-playbook <wc|sc> ../playbooks/cluster_admin_rbac.yml -b