From Theory to Terminal: A Practical Guide to Running LLMs on GCP
On September 30th GDG Cloud Budapest brings you the insight to the building blocks of Large Language Models (LLMs) again. Join us online to learn more about using offline models on GCP! ☁️
Attendance is free of charge but registration is required
💡 Abstract
Moving beyond public APIs, this session explores the practicalities of running powerful open-source Large Language Models (LLMs) within your own Google Cloud environment. We will demystify the business case for a private AI strategy, focusing on how hosting models in your GCP project enhances data security, provides predictable costs, and grants ultimate control. From selecting the right Compute Engine instances to leveraging the experimentation tools in Vertex AI, you will gain a clear roadmap for efficient deployment. We will then demonstrate how to programmatically interact with these models and use Retrieval-Augmented Generation (RAG) to securely query your internal documents.
Attendees will leave with the confidence to deploy their first private, customized LLM using Google Cloud's powerful infrastructure.
👨💻 Suggested audience: Architects and IT experts who are not specialized on the AI technologies, but they are aware of the basic terminology of AI, and are interested in the technology of AI. We are building on the knowledge we shared at the first episode of the series Technology behind the Magic of AI: Introduction, however we repeat the most important elements of the first episode as a refresher.
🗣️ Language: English
✏️ Save the date, and don’t miss out on this opportunity to learn, share, and innovate together!