Skip to content

Details

Scaling LLMs to production introduces critical challenges: How do you orchestrate multi-node execution?

Optimize GPU scheduling? Achieve low-latency data transfers between nodes? This session reveals how NVIDIA Dynamo and Azure Kubernetes Service solve these challenges together.

Through a real-world e-commerce recommendation scenario, discover how open-source frameworks and managed Kubernetes unlock enterprise-grade inference performance.

Learn to harness advanced hardware like GB200 NVL72, optimize distributed workloads on AKS, and deliver cost-efficient AI at scale to turn infrastructure complexity into competitive advantage.

Sponsors

Microsoft Reactor YouTube

Microsoft Reactor YouTube

Watch past Microsoft Reactor events on-demand anytime

Microsoft Learn AI Hub

Microsoft Learn AI Hub

Learning hub for all things AI

Microsoft Copilot Hub

Microsoft Copilot Hub

Learning hub for all things Copilot

Microsoft Reactor LinkedIn

Microsoft Reactor LinkedIn

Follow Microsoft Reactor on LinkedIn

You may also like