Kubernetes Autoscaling

Kubernetes Autoscaling is a feature for scaling nodes and pods in a Kubernetes Cluster. Autoscaling tools include Cluster Autoscaler and Horizontal Pod Autoscaler (HPA). While the scaling is automated, setting it up requires human involvement. This page gathers resources about autoscaling in Kubernetes.

Table of Contents:

Below we have compiled publicly available sources from around the world that present views on Kubernetes Autoscaling.

The Container Security book by Liz Rice

Fundamental Technology Concepts that Protect Containerized Applications

Case Studies

Specific Environments

Further Reading

  • Installing Kubernetes —There are many ways to install Kubernetes Guide and the obvious starting point is the setup section, but the installation process can sometimes be a challenge. This page gathers resources about how to install Kubernetes on various environments like Ubuntu, Windows and CentOS.

  • Kubernetes Configuration — Kubernetes Guide reads YAML files to configure services, pods and replication controllers.This page gathers resources about working with the Kubernetes configuration to deploy containers.

  • Kubernetes Monitoring — Monitoring Kubernetes effectively requires to rethink and reorient all monitoring strategies, especially if using traditional hosts such as VMs or physical machines. This page gathers resources about how to monitor Kubernetes cluster with tools like Prometheus and Datadog.

  • Kubernetes Debugging and Troubleshooting — Resources about the process of managing and maintaining production-grade, highly available Kubernetes clusters, including Kubernetes security, Kubernetes networking, Kubernetes load balancing and more.

  • Kubernetes Load Balancing — Load balancing is a relatively straightforward task in many non-container environments, but it involves a bit of special handling when it comes to containers. There are two different types of load balancing in Kubernetes - Internal load balancing across containers of the same type using a label, and external load balancing. This page gathers resources about how to configure and use the Kubernetes load balancer feature.

  • Kubernetes Security — Kubernetes Guide provides many controls that can improve application security. Configuring them requires intimate knowledge with Kubernetes and the deployment’s security requirements. This page gathers resources about security best practices for Kubernetes, including best practices for deployment, sharing data and network security.

  • Kubernetes Networking — Kubernetes does not provide any default network implementation, rather it only defines the model and leaves to other tools to implement it. There are many implementations nowadays like Flannel, Calico and Weave. This page gathers resources about how to set up highly available networked Kubernetes clusters.

  • Kubernetes Storage Management — Kubernetes is complex, and difficult to set up and configure. The best approach, therefore, is to seek out a complete container solution that includes Kubernetes as a supported, maintained component. This page gathers resources about the leading Kubetnetes distributions such as OpenShift, Canonical, Rancher and more.

  • Kubernetes in Production — The default configurations for Kubernetes Guide components are not designed for heavy and dynamic production workloads, characteristic of DevOps environments and micro-services based application deployments where containers are quickly created and destroyed. This page gathers resources about how to create a production-ready Kubernetes cluster, including examples and tutorials.

  • Working with Kubernetes Ingress — Kubernetes ingress is a collection of routing rules that govern how external users access services running in a Kubernetes cluster.  This page will introduce general strategies in Kubernetes for ingress, tutorials on how to build and troubleshoot Kubernetes Ingress controller and more.

Get updates on container technology