Deploying ML Models at Scale: Autoscaling with Ray.io and Kuberay


Details
Welcome Pythonistas!
This month, Vishnu Vardhan will talk about building ML pipelines for your Pythonhon applications and how to dynamicallylly scale them to fit your needs.
Talk Description:
In the realm of large-scale machine learning (ML) deployments, the ability to dynamically scale resources to meet varying workloads is essential for optimizing performance and efficiency. This presentation delves into the strategies and technologies for deploying ML models at scale, with a specific focus on autoscaling using Ray.io and Kuberay.
We will start by introducing Ray.io, a powerful distributed computing framework designed to simplify the scaling of Python-based applications. Attendees will learn how Ray.io can be utilized to enhance the scalability and efficiency of ML model deployments, leveraging its robust scheduling and task execution capabilities.
Building on this, we will explore Kuberay, an open-source project that seamlessly integrates Ray.io with Kubernetes, enabling efficient autoscaling of ML workloads. We will discuss the architecture and features of Kuberay, demonstrating how it facilitates the deployment and management of large-scale ML models in production environments.
By the end of this presentation, attendees will have a comprehensive understanding of how these technologies can be used to optimize their ML operations and drive innovation in their organizations.
We meet monthly for good discussion and Python shenanigans. You can show off a project you're working on or any problems that we can help solve. We're always looking for people to give lightning, beginner, and skill-based talks. Message us if you're interested in speaking!
You can watch our past meetings at watch.pyorl.org
See you all there 😃

Every 4th Tuesday of the month
Deploying ML Models at Scale: Autoscaling with Ray.io and Kuberay