Load Balancer for Managed Kubernetes Service

Load Balancer OVHcloud

Manage variable traffic loads on your containerised application

Access reserved for users of the Managed Kubernetes Service.

As your business grows and your application experiences more varied traffic, it is vital to maintain the same level of service. This is why cloud applications are usually built on distributed architectures that are spread out. They are more robust, and can easily handle peak loads. With our Load Balancer, you can securely and automatically balance your application’s load in real time, across several nodes.

99.99% availability

The OVHcloud Load Balancer is designed to deliver a high level of availability and resilience, and is also based on its own distributed architecture.

Automated node management

If a node stops working properly, it is automatically removed from the list of available nodes for balancing. This means you can easily manage maintenance operations, preventing downtime.

Directly integrated into Kubernetes

The Load Balancer delivers an interface that is directly compatible with Kubernetes. This means you can easily control your Load Balancer, with native tools.

ISO/IEC 27001, 27701 and health data hosting compliance

Our cloud infrastructures and services are ISO/IEC 27001, 27017, 27018 and 27701 certified. Thanks to our compliance, you can host healthcare data securely.

Use cases

Web applications with high volumes of traffic

Is your application generating high volumes of traffic, with an increasing number of visits? With the OVHcloud Load Balancer solution, you can manage this growth seamlessly by adding new nodes to your configuration in just a few clicks.

Variable or seasonal activity

Information websites and online stores can experience very quick variations in traffic volume. Whether it is increasing or decreasing, the Load Balancer will adapt how it distributes traffic.

Upgrades applied with no interruptions

With the rolling upgrade system, your application nodes are upgraded without any interruptions, because of the Load Balancer service’s reactivity. Nodes will be automatically added and removed.

Specifications

 Our Load Balancer solution is constantly being developed. Currently, the service is working with the following limits.

TCP 10,000 connections
HTTP 2,000 req/s
Bandwidth 200Mbit/s

 

Usage

For Kubernetes:

Create a Load Balancer

kubectl -f apply load_balancer.yaml

Delete a Load Balancer

kubectl delete service load-balancer

Technical guides

Functionality

Immediate interaction

Create a Load Balancer in less than one minute, and update it almost instantly. This means you can be well-prepared for managing traffic spikes.

Kubernetes interface

Create and manage your Load Balancer via Kubernetes.

Proxy protocol

To retain the sender’s initial address, the Load Balancer integrates Proxy Protocol. This means you can perform essential actions on the nodes such as IP address filtering, generating statistics, and analysing logs.

IP address filtering

You can choose a filtering access policy by default, and provide a restricted list of IP addresses that can connect to your solution.

Connection to private networks

To keep your application nodes isolated on the private network, the Load Balancer can be used as a pathway between public addressing and your private networks, with the OVHcloud vRack.

Pricing Public Cloud

Load Balancer billing

The Load Balancer service is billed on an hourly basis.

FAQ

What is load balancing in the cloud?

Load balancing is an operation that distributes the workload among several elements capable of performing the required task. In the cloud, load balancing is most often used for network connections that correspond to the load. These network connections are also known as service requests.

How does a load balancer work?

Load balancing follows rules set up by the operator. A flat or weighted distribution is most often selected when dealing with network connections alone. For example, when considering an application distribution, you can choose one according to routing rules depending on the content served, or user identification.