
What we’re about
The group's aim is to enable people to create and deploy their own Deep Learning models built using TensorFlow, PyTorch and JAX. We will look at a combination of both beginner and intermediate models to learn key skills - as well as going to the next level with presentations about cutting edge papers and techniques.
Ideally, each event will contain something for every level of experience - without having to explain neural networks 3-times over - and present models and code that you can play around with. Essentially, we plan on simplifying Deep Learning & other types of Machine Learning into usable skills for the real world.
Last, but not least, we're very proud to have the venue sponsored by Google.
Upcoming events (1)
See all- LLMs in production : Inference, RAG, and video modelsRakuten Event Space (Raffles Place MRT), Singapore
This month is a bit different!
- LOCATION : Near Raffles Place
- SIGN-UP (including Waitlist) : https://luma.com/heqr1dt6
- NB : NO SIGN-UP VIA MEETUP, NO WAITLIST VIA MEETUP
- FOOD : None
*HELP WANTED*
MLSG needs a few volunteers to help with logistics (like checking people into the event). If you're willing to help, and want to give back to the community, please contact Martin : by emailing my first name at reddragon.aiTalks:
"Efficient LLM Fine-tuning for Semantic Search" - Dongzhe Wang
In this talk, Dongzhe will explore how large language models (LLMs) can be efficiently fine-tuned to power semantic search systemsm, using parameter-efficient fine-tuning techniques. Attendees will learn about how LLMs are being applied to improve search in practice, along with the challenges of balancing quality and latency in real-world applications. Dongzhe is a Principal Research Scientist at Rakuten Asia, and obtained a Ph.D. from Nanyang Technological University before working at Shopee and Zhuiyi.
"Efficient Inference and Serving of LLMs and Large Video-Generative Models" - Jonathan Zhao
Jonathan will explore techniques for efficiently serving LLMs and large video generative models. The session will cover methods to optimize inference performance alongside system-level strategies for scalable deployment, highlighting key differences in serving the two model types in practice. Attendees will gain insights into approaches for improving inference and serving, considering balancing quality with latency and other real-world challenges. Jonathan is a Senior Software Engineer at Rakuten, having previously done AI product development at a startup.
"IPhO Gold using Agentic Gemini" - Martin Andrews
A recent paper showed that Gemini Pro 2.5, when driven in an agentic loop, could achieve Gold medal standard on the International Physics Olympiad theory questions. This comes hot on the heels of Google's internal version getting to Gold on the IMO (Mathematics). Martin will briefly talk about the IPhO, how the agentic system works, and show some of the actual questions (so you can see what's involved).
---
Talks will start at 7:00pm and end at around 8:45pm, at which point people normally come up to the front for a bit of a chat with each other, and the speakers.
As always, we're actively looking for more speakers - both '30 minutes long-form', and lightning talks. For the lightning talks, we welcome folks to come and talk about something cool they've done with keras, PyTorch, JAX and/or Deep Learning for 5-10mins (so, if you have slides, then #max=10). We believe that the key ingredient for the success of a Lightning Talk is simply the cool/interesting factor. It doesn't matter whether you're an expert or an enthusiastic beginner: Given the responses we have had to previous talks, we're sure there are lots of people who would love to hear what you've been playing with. If you're interested in talking, please just introduce yourself to Martin at one of the events.