Skip to content

Local Learning Environment for CloudNativePG Using Docker and Kind

License

Notifications You must be signed in to change notification settings

cloudnative-pg/cnpg-playground

Repository files navigation

CloudNativePG

Local Learning Environment for CloudNativePG

Welcome to cnpg-playground, a local learning environment designed for learning and experimenting with CloudNativePG using Docker and Kind.

Prerequisites

Ensure you have the latest available versions of the following tools installed on a Unix-based system:

You don’t need superuser privileges to run the scripts, but elevated permissions may be required to install the prerequisites.

Additional Tools

For an improved experience with the CNPG Playground, it’s recommended to install the following tools:

  • curl: Command-line tool for data transfer.
  • jq: JSON processor for handling API outputs.
  • stern: Multi-pod log tailing tool.
  • kubectx: Kubernetes context switcher.

Recommended kubectl plugins:

These tools streamline working with the CNPG Playground.

Local Environment Overview

This environment emulates a two-region infrastructure (EU and US), with each region containing:

  • An object storage service powered by MinIO containers

  • A Kubernetes cluster, deployed using Kind, consisting of:

    • One control plane node
    • One node for infrastructure components
    • One node for applications
    • Three nodes dedicated to PostgreSQL

The architecture is illustrated in the diagram below:

Local Environment Architecture

Setting Up the Learning Environment

To set up the environment, simply run the following script:

./scripts/setup.sh

Connecting to the Kubernetes Clusters

To configure and interact with both Kubernetes clusters during the learning process, you will need to connect to them.

The setup script provides detailed instructions for accessing the clusters. If you need to view the connection details again after the setup, you can retrieve them by running:

./scripts/info.sh

Inspecting Nodes in a Kubernetes Cluster

To inspect the nodes in a Kubernetes cluster, you can use the following command:

kubectl get nodes

For example, when connected to the k8s-eu cluster, this command will display output similar to:

NAME                   STATUS   ROLES           AGE     VERSION
k8s-eu-control-plane   Ready    control-plane   10m     v1.32.0
k8s-eu-worker          Ready    infra           9m58s   v1.32.0
k8s-eu-worker2         Ready    app             9m58s   v1.32.0
k8s-eu-worker3         Ready    postgres        9m58s   v1.32.0
k8s-eu-worker4         Ready    postgres        9m58s   v1.32.0
k8s-eu-worker5         Ready    postgres        9m58s   v1.32.0

In this example:

  • The control plane node (k8s-eu-control-plane) manages the cluster.
  • Worker nodes have different roles, such as infra for infrastructure, app for application workloads, and postgres for PostgreSQL databases. Each node runs Kubernetes version v1.32.0.

Demonstration with CNPG Playground

The CNPG Playground provides an excellent environment for demonstrating the CloudNativePG operator and the broader concept of running PostgreSQL on Kubernetes. You can easily create your own scenarios and customized demo environments.

To get you started, we’ve designed a demo scenario that highlights the distributed topology feature. This scenario walks you through setting up a PostgreSQL cluster distributed across two distinct regions within the playground.

For detailed instructions and resources, visit the demo folder.

Installing CloudNativePG on the Control Plane

If you plan to use the CNPG Playground without the demo mentioned earlier, you’ll need to install CloudNativePG manually.

To install the latest stable version of the CloudNativePG operator on the control plane nodes in both Kubernetes clusters, execute the following commands:

for region in eu us; do
   kubectl cnpg install generate --control-plane | \
      kubectl --context kind-k8s-${region} apply -f - --server-side

   kubectl --context kind-k8s-${region} rollout status deployment \
      -n cnpg-system cnpg-controller-manager
done

These commands will deploy the CloudNativePG operator with server-side apply on both the kind-k8s-eu and kind-k8s-us clusters.

Ensure that you have the latest version of the cnpg plugin installed on your local machine.

Cleaning up the Learning Environment

When you're ready to clean up and remove all resources from the learning environment, run the following script to tear down the containers and associated resources:

./scripts/teardown.sh

This will safely destroy all running containers and return your environment to its initial state.

Nix Flakes

Do you use Nix flakes? If you do, this package have a configured dev shell that can be used with:

nix develop .

Using Linux or WSL2

You may need:

sudo sysctl fs.inotify.max_user_watches=524288
sudo sysctl fs.inotify.max_user_instances=512

More information in the relative ticket comment.

Using Rancher Desktop

You may need to follow the instructions in the Rancher Desktop Guide to increase the open file limit.

provision:
- mode: system
  script: |
    #!/bin/sh
    cat <<'EOF' > /etc/security/limits.d/rancher-desktop.conf
    * soft     nofile         82920
    * hard     nofile         82920
    EOF
    sysctl -w vm.max_map_count=262144
    sysctl fs.inotify.max_user_watches=524288
    sysctl fs.inotify.max_user_instances=512

About

Local Learning Environment for CloudNativePG Using Docker and Kind

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •