get started deploy kapsule
Jump toUpdate content

Getting started with Kubernetes Part 2 - Deploying an app with Kapsule

Reviewed on 25 September 2023 • Published on 05 May 2022
  • Kubernetes
  • k8s
  • managed-kubernetes
  • nodes
  • pods
  • pools
  • kapsule
  • image-pull-secret
  • nodeport
  • kubectl

This tutorial accompanies the second video demonstration in our series to help users get started with Kubernetes. We walk you through Kubernetes fundamentals for beginners. In this installment, we show you how to deploy a containerized application with the Scaleway Kubernetes Kapsule.

First, we review some key Kubernetes terminology (include pools, nodes, and pods) and then demonstrate how to create a Kubernetes Kapsule via the Scaleway console. Next, we show you how to install kubectl so you can connect to your cluster from the command line of your local machine, and how to create an image pull secret for your cluster.

We then demonstrate how to deploy the containerized application (via the whoami image that we created in the first video/tutorial) to our Kapsule cluster. Finally, we show how to use the Kubernetes NodePort service to expose a port, so we can test that the application is running at its endpoint.

Future videos will cover topics like load balancing and storage for your Kubernetes application.

Security & Identity (IAM):

You may need certain IAM permissions to carry out some actions described on this page. This means:

  • you are the Owner of the Scaleway Organization in which the actions will be carried out, or
  • you are an IAM user of the Organization, with a policy granting you the necessary permission sets

Why do we need Kubernetes?

In our previous tutorial, we saw how to containerize an application. We achieved this by using Docker, an open source platform for packaging applications into containers. We created and ran our containerized application on our local machine, and then pushed the container image to a Container Registry.

While manually running and managing one container image on one local machine is fine, in a production environment we might need huge amounts of containers running simultaneously on multiple machines. It is very difficult to manage this by hand, and that is where Kubernetes can help us!

Key concepts: clusters, nodes, pods and more

Before starting the practical steps of this tutorial, we review a few key Kubernetes concepts that must be understood first:

  • Kubernetes: An open-source container orchestration platform, which automates the deployment, management, and scaling of containerized applications.

  • Cluster: If you’re running Kubernetes, you’re running a cluster. A cluster is a set of machines, called nodes, running containerized applications managed by Kubernetes. A cluster has several worker nodes and at least one Control Plane.


  • Node: A node may be a virtual or physical machine. In the case of Scaleway Kubernetes Kapsule, a node is an Instance.

  • Worker node: Each worker node runs one or more containerized applications, which are grouped into Pods. Each worker node is managed by the control plane and contains the services necessary to run Pods.

    Worker node:

  • Pod: A pod is a bundle of containers working closely together. Containers grouped together in Pods might share resources such as storage and networking.

  • Kubelet: The kubelet is the primary “node agent” that runs on each worker node. It ensures that its containers are running and healthy, and can register the node with the Control Plane’s apiserver. Scaleway’s Kubernetes Kapsule manages the kubelets for you.

  • Control plane: The Control Plane, previously known as the master node, controls and manages the worker nodes and their pods in the cluster. Scaleway’s Kubernetes Kapsule takes care of all management of the Control Plane for you. The Control Plane consists of components such as: the kube api server, which acts as the Control Plane’s front end, the kube scheduler, which assigns pods to nodes, etcd, which is a consistent and highly-available key value store to back cluster data, and more.

    Control plane:

  • Pool: A pool is a group of worker nodes within the cluster that all have the same configuration. In the case of Scaleway Kubernetes Kapsule, all the nodes in the same pool will be of the same Instance type. You can add or remove pools of nodes to increase/decrease the computational power of your cluster.


  • kubectl: kubectl is the command line interface for running commands against Kubernetes clusters. You can use kubectl to connect to and manage your clusters from the command line.

  • Managed Kubernetes: Managed Kubernetes is when third-party providers take over some of the responsibility for setting up and operating Kubernetes. Scaleway has two managed Kubernetes products: Kubernetes Kapsule and Kubernetes Kosmos. Kapsule is our classic Kubernetes offer, where all nodes and their pods in the cluster are managed by Scaleway and hosted on Scaleway Instances. Kosmos allows you to have multi-cloud clusters, with pods hosted on nodes from different cloud providers.

In both cases, Scaleway walks you through the setup of your cluster and manages your Control Plane for free. All the Control Plane’s components, and kubelets on the nodes, are fully managed. There is no need to connect to your nodes directly, as all actions and configurations can be done from the Scaleway console, via the kubectl command, the Scaleway Kapsule API or via Terraform. You can also monitor your cluster from the Kubernetes dashboard web interface.

Creating a Kubernetes Kapsule cluster

The first step in our tutorial is to create a Kubernetes Kapsule cluster. This can be achieved from the Scaleway console. Follow our dedicated how-to on creating a cluster, making sure to select Kapsule instead of Kosmos. You can leave all other settings at their default values.

Installing kubectl and connecting to your cluster

The next step is to install kubectl on your local machine, and configure it to connect to your cluster. To do this, follow our dedicated how-to on connecting to a cluster with kubectl.

Create an image pull secret

In our previous tutorial, we saw how to containerize an application and create a Docker image of the application which we pushed to our Scaleway Container Registry. Logically, we next need to deploy that application image to our Kubernetes cluster. However, Kubernetes itself must be able to pull the image to its nodes. Generally, application images are in private container registries, and so Kubernetes needs to be given access in order to pull. This access is given via image pull secrets. You might also hear these called “Docker secrets”, or just “secrets”.

  1. Go to the command line of your local machine, and enter the following command:

    kubectl create secret docker-registry registry-secret --docker-server=rg.fr-par.scw.cloud --docker-username=my-namespace --docker-password=$SCW_SECRET_KEY

    Make sure you replace rg.fr-par.scw.cloud with the endpoint of the container registry where you pushed the whoami image from the first tutorial, and my-namespace with the relevant container registry namespace. You should also have created an API key and exported it as an environment variable called $SCW_SECRET_KEY.

  2. Run the following command to display the generated secret and check that everything went well.

    kubectl get secret registry-secret --output=yaml

Deploying an application to the cluster

Now that we have created our cluster, connected to it with kubectl and defined an image pull secret so that our cluster can pull the image it needs, we can define our deployment.

  1. Create a file called whoami-deployment.yaml on your local machine:

    nano whoami-deployment.yaml
  2. Paste the following code into the file, then save and exit:

    apiVersion: apps/v1
    kind: Deployment
    name: mydeployment
    app: mydeployment
    replicas: 2
    app: mydeployment
    app: mydeployment
    - name: mycontainer
    image: rg.fr-par.scw.cloud/videodemo/whoami:latest
    - name: registry-secret

    Be sure to replace rg.fr-par.scw.cloud/videodemo/whoami:latest with the relevant path to where your whoami image is stored in your container registry.


    To better understand the different parts of the deployment yaml, here is some further information:

    • apiVersion specifies which version of the Kubernetes API to use to create the object
    • kind specifies the kind of object defined in this yaml file, here a deployment. Other Kubernetes objects we can define in yaml files include Pods, Secrets, Namespaces and more.
    • metadata helps uniquely identify our deployment object: we give it a name and a label.
    • spec shows the desired state for the deployment (2 replicas) and the selector field defines how the Deployment finds which Pods to manage, in this case using our predefined label
    • template labels the Pods with that same label, and tells them which container to run, and where to find the image for it. It also gives the pull secret needed for access. As this is a very basic yaml file for a Kubernetes deployment, you could reuse the code and adapt it for other simple deployments by changing the registry and image information.
  3. Run the following command to deploy the deployment on the cluster:

    kubectl apply -f whoami-deployment.yaml
  4. Run the following command to list all the cluster’s resources. If everything went well with the deployment, you should see that your pods are running:

    kubectl get all
  5. Run the following command, replacing the pod ID with an appropriate output from the previous command, to check that the application is running and listening on a port:

    kubectl logs pod/mydeployment-5599cbcb56-x6lb8

Exposing the service for testing via NodePort (optional)

Our application is up and running, and we could just stop at this point. However, we can carry on and expose the port it is running on, so we can access the cluster’s endpoint and check if the application is printing out its container ID as it should. We will achieve this via NodePort, a Kubernetes service which opens a port on every node of the cluster. Any traffic the cluster receives on this node is forwarded.

  1. Run the following command to create a NodePort service. Replace 80 with whatever port your pod said it was listening on at the end of the previous section.

    kubectl expose deployment mydeployment --type NodePort --port 80
  2. Use the following command to check that the Nodeport service is up and running:

    kubectl get svc

    You should see an output similar to the following:

    service/kubernetes ClusterIP <none> 443/TCP 68d
    service/mydeployment NodePort <none> 80:30564/TCP 6s
  3. Run the get nodes command to view extra information about each of your cluster’s nodes:

    kubectl get nodes -o wide

    In the output, one of the values you should see is your cluster’s external IP:

  4. Enter the following address in a browser. Replace <external-IP> with your cluster’s external IP, and <port> with the port that the NodePort service showed it was listening on in step 2 (e.g. 30564):


You should see that the whoami application is printing the ID of the container it is running from within your cluster.