Skip to content

Productionalizing ML at Scale with Ray Serve

Photo of Jules Damji
Hosted By
Jules D.
 Productionalizing ML at Scale with Ray Serve

Details

Welcome to our Ray meetup, where we focus on Ray and Ray’s native libraries for scaling machine learning workloads. We will discuss Ray Serve, an ML framework-agnostic, production-ready, operational, and scalable model serving library for this meetup.

The talks will cover three functional areas of model serving:

  • An overview of Ray Serve features and functionality and roadmap
  • On building multi-model inference graphs with Ray Serve and scaling with Ray
  • Operationalizing Ray Serve

To tie all these aspects of Ray Serve together, we will show a demo.

Agenda:

  • 6:00 PM Welcome remarks, Announcements & Agenda by Jules Damji, Anyscale
  • 6:05 PM "Ray Serve: Overview and roadmap," Edward Oakes, Simon Mo Anyscale
  • 6:15 PM Q & A
  • 6:20 PM "Developing and deploying scalable multi-model inference graphs," Jiao Dong, Anyscale
  • 7:00 PM Q & A
  • 7:10 PM "Operationalizing Ray Serve," Shreyas Krishnaswamy, Anyscale
  • 7:40 PM Q & A
  • 7:45 PM Demo Q & A

Join us if you are interested in serving and operationalizing ML models at scale using Ray Serve!

Talk 1: Ray Serve: Overview and future roadmap

In this introductory session, we’ll discuss the motivation behind Ray Serve, who’s using Ray Serve and why, recent features and updates, and look to the future feature roadmap as we approach Ray 2.0.

Bio: Edward Oakes is a software engineer and project lead on the Ray Serve team. He works across the stack at Anyscale, from Ray core to Ray Serve to the Anyscale platform.

Talk 2: Developing and deploying scalable multi-model inference graphs

In this talk, we aim to show how to leverage the programmable and general-purpose distributed computing ability of Ray to facilitate the authoring, orchestrating, scaling, and deployment of complex serving graphs as a DAG under one set of APIs, like a microservice, so a user can program multiple models dynamically on your laptop as if you’re writing a local python script, deploy to production at scale, and upgrade individually.

Bio: Jiao Dong is a software engineer focusing on Ray Serve and Ray infrastructure at Anyscale.

Talk 3: Operationalizing Ray Serve

In this session, we will introduce you to a new declarative REST API for Ray Serve, which allows you to configure and update your Ray Serve applications without modifying application files. You can incorporate this API into your existing CI/CD process to manage applications on Ray Serve as part of your MLOps lifecycle.

Bio: Shreyas Krishnaswamy is a software engineer focusing on Ray Serve and Ray infrastructure at Anyscale.

Photo of Ray Meetup group
Ray Meetup
See more events
Online event
This event has passed