Personal and Professional blog
by Dan Pilch
When setting up Kubernetes health checks with an AWS Application Load Balancer (ALB), we ran into a few challenges and had to refine our approach through trial and error. Here’s what we learned and the best practice we landed on.
Kubernetes provides two (now three) types of health checks:
200 OK
unless the container is in a fatal state and needs to be restarted.200 OK
only when the application is fully ready to serve traffic.How these interact with an ALB is critical to ensure smooth deployments and graceful shutdowns.
We configure the ALB to check only the Readiness probe. This ensures that traffic is routed only to pods that are ready to serve requests.
Here’s the typical lifecycle of our application running inside a Kubernetes pod:
200 OK
– This tells Kubernetes that the container is running.200 OK
.Now, let’s look at what happens when a pod needs to shut down.
preStop
hook.SIGUSR1
to the application and waits for N
seconds.SIGUSR1
by setting the Readiness probe to fail.preStop
hook completes its wait time and returns.SIGTERM
to the pod.This approach ensures that the ALB never sends traffic to a pod that is in the process of shutting down.
Beyond startup and shutdown, the Readiness probe can also act as a signal for auto-scaling. If your application becomes overloaded, it can temporarily fail the Readiness check to indicate that it shouldn’t receive more traffic. This can serve as an additional metric for scaling decisions.
By aligning Kubernetes health checks with ALB behavior, we’ve been able to achieve smoother deployments, zero-downtime updates, and more reliable scaling. Hopefully, this saves you from some of the pitfalls we encountered along the way!
tags: