What is Kubernetes?
We take a look at the open source platform powering containerisation at scale
First developed by Google, Kubernetes – pronounced koo-ber-net-ees – is a platform grounded in open source principles that’s designed to help manage Linux-based containerised services and workloads. This open source project, which automates application deployment, was initially designed by Google in 2014, before it branched off into a separate entity managed by the Cloud Native Computing Foundation.
However, before you try to get to grips with Kubernetes, It’s crucial to understand containerisation first. Containerisation, the process of running apps and services in isolated environments, may sound like a straightforward concept, but the underlying processes render this a much more complex undertaking.
The process of containerisation involves placing all the elements that create an app – from runtimes, config files, libraries and runtimes – in one isolated environment, known as a container. Since all the dependencies are in a single location, the container itself can be taken and moved from location to location without anything being affected. The container, for example, can be moved from an on-prem to a cloud environment, and the other way around, without all the compatibility and performance headaches that would normally arise.
The true power of containers is that they can be linked up together to create something greater, even if these containers aren’t in the exact same location. This is because these entities can communicate with one another across environments to create a complete application without having to employ a single virtualised environment or operating system.
This is an increasingly popular form of software deployment, particularly in recent years, but has also proven itself to be increasingly complex, especially with businesses that wish to deploy multiple containers across several machines – both physical and virtual machines. Manual processing may be required, as well as continuous management that deploying multiple containers demands.
This may not be such a significant barrier when engaging in containerisation on a simple level, but as development scales-up, several containerised applications may be needed to work in tandem to power a business’ services. When containerisation becomes this complex, the number of containers may grow exponentially and become impossible to manage.
How to maximise the value of your data and apps with IaaS
Free yourself from infrastructure complexityDownload now
Kubernetes seeks to eliminate this. Originally developed by a team at Google, a company that today has everything running in containers, Kubernetes acts as an orchestration tool, giving users an overview of their container deployments. This makes it far easier to operate generally as well as making it possible to have hybrid, public and private cloud containers running simultaneously.
Kubernetes has a bunch of tools that make all of this possible, including the option to sort containers into groups, or 'pods', which then makes it easier to serve the applications with the necessary infrastructure, such as storage and networking capabilities. It handles a lot of the optimisation work so that businesses can focus on what they want their services to achieve, rather than the worry about whether apps are talking to each other.
It's also able to optimise your hardware to ensure the correct amount of resources are being applied to each application, and add or remove resources depending on whether you want to scale up or down. Automated health checks also mean that errors can be corrected without human intervention, and it also has provisions to roll out updates to containers without downtime.
Perhaps the most important thing is that Kubernetes is not tied to a specific environment and it can operate regardless of where your containers are, whether that's in a public cloud, private, a virtualised system, or even a single laptop, and even combine all of these together.
Speaking at the recent VMworld conference in California, VMware CEO Pat Gelsinger took time to highlight Kubernetes as “the de facto API for today's multi-cloud world”.
“Much like Java, two decades ago, Kubernetes is a rare technology that brings everyone together,” he said.
Who owns Kubernetes?
Google would eventually donate the Kubernetes platform to the Cloud Native Computing Foundation in 2015, releasing it into the open source community to be used freely by anyone.
Although it primarily works with Docker, a program that builds containers, Kubernetes will work with any platform that conforms to the Open Container Initiative (OCI) standards that define container formats. (Note: Docker has some higher-level orchestration tools that essentially perform the same functions as Kubernetes).
As Kubernetes is an open-source technology, there's no single service available with dedicated support. The technology has essentially been adapted by various vendors into their own flavours, whether that's Google, Amazon Web Services, or Red Hat, and choosing one will depend on the services you currently use, or want as part of a contract.
Other providers include Docker, Canonical, CoreOS, Mirantis, and Rancher Labs. The latter was recently acquired by German-based Linux distribution company SUSE in a deal thought to be worth between $600 million to $700 million.
Rancher Labs, founded in 2014 and currently employing more than 200 people, provides open-source software that allows organisations to deploy and manage Kubernetes at scale.
The Cupertino-based startup claims to be the "most widely used enterprise Kubernetes platform", boasting 30,000 active users. It's customer base includes American Express, Comcast, Deutsche Bahn and Viasat.
The Kubernetes language
In order to fully understand Kubernetes, you need to learn the vernacular that comes with it.
Each deployment follows the same basic hierarchy: Cluster > Master > Nodes > Pods
Let's start at the top. Kubernetes is deployed in a 'cluster' – this is a collective term referring to both the group of machines that are running the platform and the containers that are managed by them.
Within each cluster there are multiple 'nodes' – these are normally the machines that the containers are running on, whether that's virtualised or physical, and multiple containers may be hosted on a single node (with each container hosting an application).
Each 'cluster' must always have a 'master', which acts like a management window from which admins can interact with the cluster. This includes scheduling and deploying new containers within the nodes.
Nodes are responsible for creating 'pods' – the term given to an instance of an application that's running within the cluster, usually involving multiple containers. This means that users are able to visualise all the individual containers supporting an application as a single entity.
Pods can be best thought of as the basic building block within Kubernetes, and are created based on the needs of the user.
Demand for Kubernetes skills
In the past few years, containerisation has become more and more popular within app deployment, a trend that is also mirrored in the job market. In 2018, the demand for developers and engineers with experience in Kubernetes reached new heights, when IT Jobs Watch registered an almost eight-fold increase in these kinds of roles in a mere two years. When this happened, Josh Kirkwood, CyberArk DevOps Security Lead, said that Kubernetes had “become a massive money word”. He added that “these figures show that DevOps teams are seeking more skills to help them manage and deploy applications at scale”.
By now, even though many may have forgotten that Kubernetes has only been around for seven years, it has become a staple in the DevOps industry. Last year, IBM made headlines when it posted a job advert requiring a “minimum” of 12 years experience in Kubernetes, which included deploying microservices and other platforms, “hands-on” experience setting up platforms and managing secure secure secrets, as well as knowledge of container orchestration. According to several Twitter users, the requirements for the role were rather outlandish, especially since the very first GitHub post about Kubernetes originated on 7 June 2014.
It has become more accessible to learn and gain experience with Kubernetes, which has been more available following the increase in demand for skills with the system. Google Cloud announced earlier this year it would offer free training for AI, multi-cloud services, machine learning, and data analytics, which included routes to foundational certificates. One course asks learners to demonstrate core infrastructure skills, such as how to deploy a virtual machine, write cloud shell commands, and run applications on Kubernetes.
Next-generation time series: Forecasting for the real world, not the ideal world
Solve time series problems with AIFree download
The future of productivity
Driving your business forward with Microsoft Office 365Free download
How to plan for endpoint security against ever-evolving cyber threats
Safeguard your devices, data, and reputationFree download
A quantitative comparison of UPS monitoring and servicing approaches across edge environments
Effective UPS fleet managementFree download