How to use Kubernetes Horizontal Pod Autoscaler

How to use Kubernetes Horizontal Pod Autoscaler

·

4 min read

What is Autoscaling and Why it is important?

Autoscaling means dynamically allocating cluster resources such as CPU and memory to your application based on real time demand. This ensures that your application is having the right amount of resources to handle different levels of load which directly improves application performance and availability.

In kubernetes, you have three main types of Autoscaler to choose from:

  • Horizontal Pod Autoscaler (HPA)

  • Vertical Pod Autoscaler (VPA)

  • Kubernetes Event Driven Autoscaler (KEDA)

In this guide we will discuss about Horizontal Pod Autoscaler (HPA)

Horizontal Pod Autoscaler (HPA)

HPA is a kubernetes feature that automatically scales the number of Pod replicas in a Deployment, ReplicaSet, or StatefulSet based on certain metrics like CPU utilization or custom metrics. Horizontal scaling is the most basic autoscaling pattern in kubernetes.

HPA sets two parameters, 1) Target Utilization level and 2) Minimum and Maximum number of replicas allowed. When the utilization of a Pod exceeds the target, HPA will automatically scale up the number of replicas to handle the increased load. Similarly when the usage drops, HPA will scale down the number of replicas to conserve resources.

When to use HPA ?

  • Scalability based on workload: HPA is ideal when you want to scale the number of replicas up or down based on the workload or incoming traffic.

  • Resource utilization scaling: HPA is useful for scenarios where you want to ensure CPU or memory utilization pods remains within a certain threshold.

Example: How to set up a new kubernetes HPA

Pre-requisites:

  • Kubectl

  • metrics-server

Steps:

  1. Create the manifest files
apiVersion: apps/v1
kind: Deployment
metadata:
  name: demo
spec:
  selector:
    matchLabels:
      app: demo
  template:
    metadata:
      labels:
        app: demo
    spec:
      containers:
        - name: demo
          image: registry.k8s.io/hpa-example:latest
          ports:
            - containerPort: 80
          resources:
            requests:
              cpu: 500m
---
apiVersion: v1
kind: Service
metadata:
  name: demo
spec:
  ports:
    - port: 80
  selector:
    app: demo

Save the file, and use kubectl apply command to apply it.

kubectl apply -f app.yaml

  1. Creating Horizontal Pod Autoscaler

The following manifest file creates the HPA and sets the minimum and maximum number of pods in order to maintain average CPU utilization of 50%

apiVersion: autoscaling/v2
kind: HorizontalPodAutoscaler
metadata:
  name: demo
spec:
  scaleTargetRef:
    apiVersion: apps/v1
    kind: Deployment
    name: my-app
  minReplicas: 3
  maxReplicas: 9
  metrics:
    - type: Resource
      resource:
        name: cpu
        target:
          type: Utilization
          averageUtilization: 50

Here's the breakdown of the components:

  • apiVersion: autoscaling/v2 - Specifies the API version for the HPA resource.

  • kind: HorizontalPodAutoscaler - Specifies the kind of resource as its an HPA.

  • metadata: Name of the HPA

  • scaleTargetRef: Deployment that the HPA will scale

  • apiVersion: This is for the deployment version.

  • kind: The kind of resource it will scale.

  • name: Name of the deployment

  • minReplicas: Atleast 3 replicas will be running

  • maxReplicas: At high demand, it will scale up to a max of 9 replicas.

  • metrics: Specifies the metrics used for autoscaling.

  • type: Resource metrics used for autoscaling.

  • resources: Specifies the resource.

  • name: Specifies CPU metric will be used.

  • target: Specifies the target.

  • type: Indicates the target is set to type utilization.

  • averageUtilization: Specifies the target utilization should be around 50%.

  1. Apply the HPA using the below command
kubectl apply -f hpa.yaml
  1. Now check the HPA status
kubectl get hpa

  1. Also check the deployment status.
kubectl get deployments

The deployment has scaled upto 3 replicas automatically. This is the same replica count which we mentioned inside the HorizontalPodAutoscaler manifest file.

  1. Generate load test.
kubectl run -it --rm load-test --image=busybox:1.28 --restart=Never -- /bin/sh -c "while sleep
0.01; do wget -q -O- http://demo; done"

Wait for a minute to see the increased load, you can observe the same using the get hpa command.

kubectl get hpa

This time the CPU is 20% and the load is increasing slowly. Now if you check again after sometime there will be more increase in the load which will significantly increase the number of pods.

In this series of events, you can see that as the load increases beyond the target utilization, the HPA will try to maintain the limit by increasing the number of pods. The replica count rises to match the resource usage. Later, when the load test is stopped, the HPA scales down the number of pods, which is exactly what we needed.

We've successfully used HPA to automate our application. HPA is powerful mechanism for scaling your application workloads.

Conclusion:

In this blog, we covered the basics of HPA and how to create an HPA using YAML configuration. The HPA reacts to changes in the metrics and adjusts the number of replicas. In the upcoming blogs we will discuss about the remaining types of Autoscaler.

If you found this post useful, give it a like👍

Repost♻️

Follow Bala for more such posts 💯🚀