Autoscaling Applications Deployed to Kubernetes
Containers and orchestration promise greater resource efficiency and simplified deployments. Developers provide code in the form of a container, define their required resources for operation, and the orchestrator does the rest. Sounds easy, but if your workloads are dynamic in nature, how can you ensure sufficient resources to meet the performance and availability requirements of your customers?
In this workshop we will demonstrate how to autoscale your application workloads on Kubernetes. You will learn frameworks for how to identify their key work and resource metrics; as well as how to use them to drive horizontal and vertical pod autoscaling so that you can maximize efficiency, while ensuring service reliability.