Ollama on Kubernetes : How to Deploy Ollama on Kubernetes

Details
Join Mike Peterson as he dives into a practical, hands-on demo showcasing how to deploy Ollama on Kubernetes using vCluster to create a cost-effective, multi-tenant environment. In this session, Mike will demonstrate how to set up virtual clusters that provide isolated Kubernetes environments running on shared infrastructure, enabling efficient GPU usage for multiple users.
Date & time: September 11, 5:30 PM
Location: Online (Teams Webinar)
Presented by: Mike Peterson, Senior Technical Marketing Engineer at vCluster.
Attendees will learn:
- How to deploy Ollama in a virtual cluster on Kubernetes
- Setting up GPU runtime classes for optimized GPU sharing
- Creating deployments, services, and ingress resources for Ollama
- Connecting and interacting with Ollama running inside the Kubernetes cluster using the Ollama client on a local machine
- Tips on multi-tenancy, managing costs, and securing access with OAuth techniques
This webinar is perfect for cloud-native developers and platform engineers interested in managing AI workloads efficiently on Kubernetes while minimizing costs. Join to see how vCluster enables scalable, secure, and cost-effective deployments of AI models with practical insights and live demonstrations.
Connect With Us on Social Media
Instagram: https://www.instagram.com/torontoai_
Facebook: https://www.facebook.com/profile.php?id=100068811921154
LinkedIn: https://www.linkedin.com/company/torontoai/
TikTok: https://www.tiktok.com/@torontoai_?_t=zs-8yvjfept7ob&_r=1

Ollama on Kubernetes : How to Deploy Ollama on Kubernetes