Vancouver, Canada - July 2025
Generative AI models are trained on internet-scale datasets, yielding powerful capabilities but also introducing risks like copyright infringement, PII leakage, and harmful knowledge. Targeted removal or unlearning of sensitive data is challenging, as retraining on curated sets is computationally expensive, driving research into machine unlearning and model editing. Yet approaches like RLHF only suppress undesirable outputs, leaving underlying knowledge vulnerable to adversarial extraction. This raises urgent privacy, security, and legal concerns, especially under the EU's GDPR "right to be forgotten". Because neural networks encode information across millions of parameters, precise deletion without degrading performance is complex, and adversarial or whitebox attacks can recover ostensibly erased data.
This workshop brings together experts in AI safety, privacy, and policy to advance robust, verifiable unlearning methods, standardized evaluation frameworks, and theoretical foundations. By achieving true erasure, we aim to ensure AI can ethically and legally forget sensitive data while preserving broader utility.
We invite contributions exploring key challenges and advancements at the intersection of machine unlearning and generative AI.
Coming soon!
Time | Event |
---|---|
09:00 AM - 09:10 AM | Opening Remarks |
09:10 AM - 09:40 AM | Sijia Liu's talk: LLM Unlearning: Opportunities and Limitations |
09:40 AM - 10:25 AM | Live Poster Session 1 |
10:25 AM - 10:45 AM | Coffee Break |
10:45 AM - 11:15 AM | Nicholas Carlini's talk: Adversarial ML: harder than ever |
11:15 AM - 11:30 AM | Contributed Talk 1: Do Not Mimic My Voice: Speaker Identity Unlearning for Zero-Shot Text-to-Speech |
11:30 AM - 11:45 AM | Contributed Talk 2: Distributional Unlearning: Forgetting Distributions, Not Just Samples |
11:45 AM - 12:00 PM | Contributed Talk 3: Unlearning Isn't Invisible: Detecting Unlearning Traces in LLMs from Model Outputs |
12:00 PM - 12:30 PM | Peter Hase's talk: Beyond Retain and Forget Sets: Unlearning as Rational Belief Revision |
12:30 PM - 01:30 PM | Lunch Break |
01:30 PM - 02:00 PM |
Eleni Triantafillou's talk: What makes unlearning fail?: An investigation into interpretable factors underlying failure modes of unlearning algorithms and how to alleviate them |
02:00 PM - 02:30 PM | Shagufta Mehnaz's talk: Robust Unlearning for Large Language Models |
02:30 PM - 03:00 PM | Ling Liu's talk: Machine Unlearning of GenAI: Bright Side vs. Dark Side |
03:00 PM - 03:45 PM | Live Poster Session 2 |
03:45 PM - 04:00 PM | Coffee Break |
04:00 PM - 04:55 PM | Live Panel Discussion with Speakers and Panelists: Nicholas Carlini, Eleni Triantafillou, A. Feder Cooper, Amy Cyphert |
04:55 PM - 05:00 PM | Closing Remarks |