Skip to content

ML Meetup: LLMs in Production and Semantic Search Engine

Photo of
Hosted By
ML Meetup: LLMs in Production and Semantic Search Engine


Welcome to our in-person ML monthly meetup in San Francisco. Join us for deep dive tech talks on AI/ML, food/drink, networking with speakers&peers developers, and win lucky draw prizes.

[RSVP instructions]


  • 5:00pm~5:30pm: Checkin, Food and Networking
  • 5:30pm~5:40pm: Welcome/Sponsor intro
  • 5:40pm~7:30pm: Tech talks
  • 7:300pm~8:00pm: Open discussion, Lucky Draw & Mixer

Tech Talk 1: LLMs in Production
Speaker: Chenggang Wu, Co-founder and CTO @Aqueduct
Abstract: Large language models (LLMs) and other foundation models have made an unprecedented step forward in AI. Unfortunately, the infrastructure required to run these models is so overwhelmingly complex that only a few select companies have the requisite capabilities. Infrastructure challenges range from managing large amount of data, deploying complex pipelines, and managing compute services. In this talk, we will discuss:
- An overview of the infrastructure challenges of running LLMs in the cloud
- A demo/walkthrough of deploying an LLM on existing cloud infrastructure
- How Aqueduct seamlessly takes a LLM-powered workflow from prototype to production.

Tech Talk 2: Open Source Semantic Search Engine
Speaker: Ken MacInnis @Weaviate
Abstract: Today, we've gotten used to natural language search and recommendation systems. We expect to get what we search for without remembering the exact keywords. To enable this, we use machine learning models that represent data as vectors that capture the semantics of the data. To do this at scale we need a way to efficiently store and retrieve large amounts of vector and non-vector data. Scaling ML models to work reliably in production is complex and implementing efficient vector search while keeping real-time CRUD support that is expected from databases is even harder. Vector search engines form an elegant solution to these challenges. A vector search engine helps you create machine-learning-first search and recommendation systems based on your data. It searches through your data super-fast with Approximate Nearest Neighbor (ANN) search, while also supporting all CRUD operations and data mutability.

In this session, you will learn what vector search is, why and when you would need it and you will see vector search in action during live demos.

Internet of Things Developers - IoTD
Internet of Things Developers - IoTD
See more events
· San Francisco, CA