KEDA, Scale Your Kubernetes Workload on Your Own Terms

Kubernetes is a powerful platform to host various kind of workloads, and these workloads vary in their need of scale. For that, Kubernetes has a built-in functionality to scale these workloads based on their resources consumption like CPU and memory. However, there is no built-in way to scale workloads based on events that happen outside of the cluster; e.g. the length of a storage queue in the cloud. KEDA came in to fill this gap with various built-in scalers that come with the package. You can also write your own scaler that responds to your own events and needs. In this session we will understand what KEDA is, how it works, and how we can build our own scaler that scales our workloads to our own events and needs.

    Emad Alashi
    Principal Consultant, Insight

    Emad Alashi is a software developer whose main interest is in software architecture, software delivery, the cloud, and the human interaction caught in between.

    He helps organisations adopt the cloud, embrace DevOps, and build modern systems that achieve the right business goals. Lately his interest has been focused on Event-Driven systems and Cloud-native

    Emad speaks regularly in conferences and user groups, including NDC Sydney, Microsoft Ignite Australia, and in local user groups and code camps like Vic.Net, Azure Meetup, Azure Bootcamps, and Alt.Net in Melbourne Australia.

    He posts to his blog on, and sometimes stream on Twitch at

    Emad, and currently works as a Principal Consultant in Insight, and can be found on twitter @emadashi.

    Programutvikling uses cookies to see how you use our website. We also have embeds from YouTube and Vimeo. How do you feel about that?