Self-Hosting AI LLMs: Deploying Ollama on Azure Container Apps
Details
Curious about running Large Language Models on your own infrastructure? Want to break free from API rate limits and usage costs? Join us for an engaging session where we dive into Ollama, the sleek, developer-friendly way to run LLMs like LLaMA 3, Mistral, and more—right on Azure!
In this talk, we’ll walk through how to containerize and self-host Ollama using Azure Container Apps. You’ll learn how to:
- Deploy container workloads with ease
- Integrate Ollama into your own apps for private, cost-efficient inference
- Secure and scale your AI services in a cloud-native way
- See a live demo of an end-to-end LLM deployment
Artificial Intelligence
Cloud Computing
Microsoft Azure
