Run LLMs Your Way: Deploying Ollama on Azure Container Apps
Details
Tired of API limits and rising inference costs? Take control of your AI infrastructure! In this session, we’ll explore how to run open-source Large Language Models like LLaMA 3, Mistral, and more using Ollama—a tool to run LLMs locally—and Azure Container Apps for scalable, secure deployments.
You’ll learn how to:
- Build and deploy Ollama in a containerized environment
- Integrate local LLM inference into your own apps
- Secure, scale (Scale down to zero when not in use!), and monitor AI workloads the cloud-native way
- Watch a live demo of an end-to-end LLM deployment on Azure
Whether you’re creating private copilots, intelligent assistants, or exploring AI self-hosting, this session will show you how to own your AI stack—from model to deployment.
🎯 Who Should Attend:
Developers, cloud engineers, architects, and AI enthusiasts ready to experiment with open-source LLMs in the cloud.
🧰 Tech Stack:
Azure Container Apps | Ollama | Docker | LLaMA 3 | Mistral | Open-Source AI
Bring your curiosity, your beverage of choice, and your sense of adventure—we’ll make the cloud talk back! 🚀


