ML Meetup: LLMs in Production and Semantic Search Engine


Details
Welcome to our in-person ML monthly meetup in San Francisco. Join us for deep dive tech talks on AI/ML, food/drink, networking with speakers&peers developers, and win lucky draw prizes.
[RSVP instructions]
- Register at the event website: https://www.aicamp.ai/event/eventdetails/W2023040417 (correct/full name is required for badges and check in. NO walk-ins, NO access without badge)
- Contact us to submit topics and/or sponsor the meetup on venue/food/swags/prizes. https://forms.gle/JkMt91CZRtoJBSFUA
- Community on Slack for events chat, speakers office hour and sharing learning resources, job openings and projects collaboration. join slack (join #sanfrancisco channel)
Agenda:
- 5:00pm~5:30pm: Checkin, Food and Networking
- 5:30pm~5:40pm: Welcome/Sponsor intro
- 5:40pm~7:30pm: Tech talks
- 7:300pm~8:00pm: Open discussion, Lucky Draw & Mixer
Tech Talk 1: LLMs in Production
Speaker: Chenggang Wu, Co-founder and CTO @Aqueduct
Abstract: Large language models (LLMs) and other foundation models have made an unprecedented step forward in AI. Unfortunately, the infrastructure required to run these models is so overwhelmingly complex that only a few select companies have the requisite capabilities. Infrastructure challenges range from managing large amount of data, deploying complex pipelines, and managing compute services. In this talk, we will discuss:
- An overview of the infrastructure challenges of running LLMs in the cloud
- A demo/walkthrough of deploying an LLM on existing cloud infrastructure
- How Aqueduct seamlessly takes a LLM-powered workflow from prototype to production.
Tech Talk 2: Open Source Semantic Search Engine
Speaker: Ken MacInnis @Weaviate
Abstract: Today, we've gotten used to natural language search and recommendation systems. We expect to get what we search for without remembering the exact keywords. To enable this, we use machine learning models that represent data as vectors that capture the semantics of the data. To do this at scale we need a way to efficiently store and retrieve large amounts of vector and non-vector data. Scaling ML models to work reliably in production is complex and implementing efficient vector search while keeping real-time CRUD support that is expected from databases is even harder. Vector search engines form an elegant solution to these challenges. A vector search engine helps you create machine-learning-first search and recommendation systems based on your data. It searches through your data super-fast with Approximate Nearest Neighbor (ANN) search, while also supporting all CRUD operations and data mutability.
In this session, you will learn what vector search is, why and when you would need it and you will see vector search in action during live demos.

ML Meetup: LLMs in Production and Semantic Search Engine