Skip to content

Details

Tired of API limits and rising inference costs? Take control of your AI infrastructure! In this session, we’ll explore how to run open-source Large Language Models like LLaMA 3, Mistral, and more using Ollama—a tool to run LLMs locally—and Azure Container Apps for scalable, secure deployments.

You’ll learn how to:

  • Build and deploy Ollama in a containerized environment
  • Integrate local LLM inference into your own apps
  • Secure, scale (Scale down to zero when not in use!), and monitor AI workloads the cloud-native way
  • Watch a live demo of an end-to-end LLM deployment on Azure

Whether you’re creating private copilots, intelligent assistants, or exploring AI self-hosting, this session will show you how to own your AI stack—from model to deployment.

🎯 Who Should Attend:
Developers, cloud engineers, architects, and AI enthusiasts ready to experiment with open-source LLMs in the cloud.

🧰 Tech Stack:
Azure Container Apps | Ollama | Docker | LLaMA 3 | Mistral | Open-Source AI

Bring your curiosity, your beverage of choice, and your sense of adventure—we’ll make the cloud talk back! 🚀

Artificial Intelligence
Containers
Microsoft Azure

Sponsors

Sponsor logo
Akumina
Akumina provides a meeting place and pizzas for our meeting.
Sponsor logo
AweMind LLC
Awemind sponsors Pizzas for our meeting.

Members are also interested in