Geshan Manandhar Geshan is a seasoned software engineer with more than a decade of software engineering experience. He has a keen interest in REST architecture, microservices, and cloud computing. He also blogs at

Intro to KubeVela: A better way to ship applications

6 min read 1680

Intro to KubeVela: A Better Way to Ship Applications

Kubernetes has already won the container orchestration battle. But the war developers still need to overcome with Kubernetes is its complexity. This is where KubeVela comes into play.

KubeVela makes deploying applications to Kubernetes much easier. Rather than knowing about service, deployment, pods, and horizontal pod scaling, you can specify a much lighter configuration. Let’s dive deeper on how.

What is KubeVela?

KubeVela is an open source tool that abstracts away the complicated Kubernetes artifacts from developers — and, to a point, platform builders — to create a simpler deployment experience.

As a software engineer, Kubernetes artifacts can be overwhelming at first. This is where KubeVela acts as a much-needed bridge that allows developers to worry less about the underlying artifacts. KubeVela helps both developers and platform builders by providing a good framework that provides a user-friendly yet highly extensible platform.

Its tagline is “Make shipping applications more enjoyable,” and I think it lives up to that slogan. KubeVela is based on the Open Application Model (OAM) specification, which puts the focus squarely on the application rather than the container or orchestrator.

Where does KubeVela fit in the Kubernetes ecosystem?

Kubernetes artifacts like service, deployment, horizontal pod scaler, etc. look daunting at first glance, but after some time, one gets used to them. There are many great tools to deploy applications to Kubernetes and abstract away the inner details of Kubernetes artifacts.

Some of these useful tools include Helm, Skaffold, and Draft, to name a few. We can find a good comparison of these types of tools in this helpful post.

Still, none of these tools are geared towards both developers and platform operators in the same way KubeVela is. KubeVela gives a user-friendly and simple way to define applications that can be easily deployed and updated on any Kubernetes cluster.

Deploy a simple Node.js express app with KubeVela

In this demo, we will deploy a simple Node.js Express application that has already been dockerized, and the Docker image is available on Docker Hub.


Before jumping in and defining the KubeVela app file, let’s first clear up some prerequisites:

  • You should know about Kubernetes and how it generally works
  • You must have kubectl and Helm installed and running on your testing environment
  • You are aware of how container registry works and how to push and pull container images

We will run this demo on the local kind Kubernetes cluster. Per its readme, kind is “a tool for running local Kubernetes clusters using Docker container ‘nodes.’ ”

Let’s go ahead and set up our local Kubernetes cluster using kind. We’ll outline the steps below.

Set up a local Kubernetes cluster with kind

Depending on your operating system, you might have a different way to install the kind CLI. I am using a Mac, so I ran the following command:

brew install kind

You will be able to find your OS-specific command to install kind on its installation page.

Before proceeding, make sure your ports 80 and 443 are open on your local machine. This means you will need to stop any servers like Apache or NGINX occupying port 80. Also, check that you don’t have any other services taking up port 80 locally like an NGINX proxy, which could be running as a Docker container, too.

To create our local Kubernetes cluster with kind, we will execute the following command:

cat <<EOF | kind create cluster --config=-
kind: Cluster
- role: control-plane
  - |
    kind: InitConfiguration
        node-labels: "ingress-ready=true"
  - containerPort: 80
    hostPort: 80
    protocol: TCP
  - containerPort: 443
    hostPort: 443
    protocol: TCP

This creates a kind Kubernetes cluster with ports 80 and 443 from the container mapping to your local machine.

It should take a couple of minutes and come back with an output like below:

The Output after Creating Our Kind Kubernetes Cluster

For the next step, we will configure our kubectl to use the Kubernetes cluster we just created by running:

kubectl cluster-info --context kind-kind

Up next, we will install Ingress in our Kubernetes cluster with kind by running:

kubectl apply -f

The above command will show us something like below:

Output After Installing Ingress

To quickly check whether we have any pods running on the newly created Kubernetes cluster, we can run:

kubectl get po -A

It will show us pods running on all namespaces like below:

Pods Running on All Namespaces

Hurray, our Kubernetes cluster is up and running. The next step is to install KubeVela — let’s go.

Install KubeVela in the local K8s cluster

Now that we have our kind Kubernetes cluster running locally inside a Docker container (try docker ps to check), we will proceed to install KubeVela. To install KubeVela with Helm, run the following command:

helm repo add kubevela && helm repo update

It will show us the following output:

Output After Installing KubeVela

To make things easier, we have clubbed together two commands. The first gets the Helm chart for KubeVela, and the second updates the help repo.

For the next step, we will again run two commands together:

kubectl create namespace vela-system && helm install -n vela-system kubevela kubevela/vela-core --set installCertManager=true

These two commands show us an output like below:

Output After Creating the vela-system Namespace

The first command here creates the vela-system namespace in our Kubernetes cluster. The second command installs the KubeVela Helm chart in the vela-system namespace created in the previous command with cert-manager.

As per the KubeVela installation documentation, KubeVela relies on cert-manager to create certificates for webhooks.

At this juncture, if we run kubectl get po -n vela-system to verify whether KubeVela is running properly, it will show us one pod whose name starts with kubevela-veal-core- running. Now we know that KubeVela is installed on our Kubernetes cluster and running successfully.

In the next step, we will install the vela CLI.

Install the KubeVela CLI

To install the vela CLI on a Mac I ran the following command:

curl -fsSl | bash

Depending on your operating system, you can find the relevant command to get the CLI on this page.

To verify that your vela CLI is installed properly, run vela version, and you should see an output like below:

Output After Running vela-version Command

We will use the vela CLI to create a vela.yaml file that defines how the app should be deployed on Kubernetes. Time to play around with the vela CLI now.

‘Velaize’ and deploy our Node.js application

Consequently, we will “Velaize” our existing Node.js express demo application that has already been dockerized. To make things simpler I have pushed the application to Docker Hub. As you might already be aware, to deploy our application, the container should be pushed to a container registry. We will be using Docker Hub as our container registry for this demo.

To try the demo app on your Kubernetes cluster, you can go to an empty folder and run:

vela init

It will ask you some questions like the name of the app, name of the service, workload type, and Docker image for the service. Fill it up like below; the main thing here is the image of the service, which points to geshan/nodejs-docker-express, hosted on Docker Hub:

Running the App on Our Kubernetes Cluster

Our Node.js application will run on port 3000, and by default, the latest tag for the Docker image will be pulled in and deployed to Kubernetes by KubeVela.

Run vela status testapp to view the status of the app as follows:

Viewing the Status of Our App

Another way to quickly check it is by running kubectl get po, which should show a pod running.

The vela init command has created a vela.yaml file with the following contents:

createTime: "2021-02-09T13:51:25.723862+11:00"
name: testapp
    image: geshan/nodejs-docker-express
    port: 3000
    type: webservice
updateTime: "2021-02-09T13:51:25.723863+11:00"

Test the Node.js app

To test the Node.js Express app running on your local K8s cluster deployed with KubeVela, run the following command:

vela port-forward testapp

It will show an output like below:

Testing Our App

Now, we can hit on the browser to see our app running in the Kubernetes cluster on the container as follows:

Viewing Our App Running Successfully on Our Kubernetes Cluster

Furthermore, when the Docker image changes and updates are pushed on the Docker image registry, we can run vela up to redeploy the application.

Deploy a change with KubeVela

For demo purposes, I am going to change the message from “Welcome to Express with Docker” to “Welcome to Express with Docker and KubeVela” and push it as a new Docker tag called kubevela. By default, until now, it was taking the latest tag.

Then, we will change the tag in the vela.yaml file, as below:

createTime: "2021-02-09T13:51:25.723862+11:00"
name: testapp
    image: geshan/nodejs-docker-express:kubevela
    port: 3000
    type: webservice
updateTime: "2021-02-09T13:51:25.723863+11:00"

The change is adding the tag to the Docker image as image: geshan/nodejs-docker-express:kubevela. Given the tag is available, we can do:

vela up
vela status testapp #confirm the app is up and healthy
vela port-forward testapp

Then, we can see the new changes on the browser at: as follows:

Viewing the Changes to Our App

There you have it! We ran a small Node.js Express app that had been previously dockerized on a local Kubernetes (kind) cluster with KubeVela. I think it was much easier than dabbling with service, deployment, and other Kubernetes artifacts.

The local Kubernetes cluster should be treated as your playground for the proof of concept. We can realize the ease and flexibility KubeVela adds when we want to deploy our containerized apps to production-grade Kubernetes clusters.

As this is an intro post, I decided not to dive deeper into the routing and auto-scaling parts. There are some alternative commands to add these capabilities, too. I would leave it up to you to explore these powers of KubeVela.


We have seen how to deploy a Node.js Express app to Kubernetes using KubeVela. KubeVela makes it easy and a lot more user-friendly to deploy any containerized application into Kubernetes. I would recommend trying KubeVela if you and your team are venturing into the Kubernetes landscape for its ease of use and flexible architecture.

Get setup with LogRocket's modern error tracking in minutes:

  1. Visit to get an app ID.
  2. Install LogRocket via NPM or script tag. LogRocket.init() must be called client-side, not server-side.
  3. $ npm i --save logrocket 

    // Code:

    import LogRocket from 'logrocket';
    Add to your HTML:

    <script src=""></script>
    <script>window.LogRocket && window.LogRocket.init('app/id');</script>
  4. (Optional) Install plugins for deeper integrations with your stack:
    • Redux middleware
    • ngrx middleware
    • Vuex plugin
Get started now
Geshan Manandhar Geshan is a seasoned software engineer with more than a decade of software engineering experience. He has a keen interest in REST architecture, microservices, and cloud computing. He also blogs at

Leave a Reply