AI Safety Thursdays: Can we make LLMs forget? An intro to machine unlearning


Details
Registration Instructions (IMPORTANT)
This is a paid event ($5 general admission, free for students) with limited tickets - you must RSVP on Luma to secure your spot.
If you can't make it in person, feel free to join the live stream at 6:30 pm, via this link.
Description
LLMs are pre-trained on a large fraction of the internet. As a result, they can regurgitate private, copyrighted, and potentially hazardous information, causing deployment and safety challenges.
Lev McKinney will guide us through machine unlearning in LLMs—how models retain facts, methods for identifying influential training data, and techniques for suppressing predictions. Finally, we'll assess current research and its effectiveness for policy and safety concerns.
Timeline
6:00 to 6:30 - Food & Networking
6:30 to 7:30 - Main Presentation & Questions
7:30 to 8:00 - Discussion

AI Safety Thursdays: Can we make LLMs forget? An intro to machine unlearning