Machine Unlearning for AI Safety
The ability to selectively remove undesirable learned information (such as private data, copyrighted content, or harmful knowledge that could facilitate the misuse of generative models) is increasingly recognized as a critical capability for trustworthy AI. This process, known as machine unlearning (MU), has become essential as generative models are deployed in sensitive domains including healthcare, defense, personalized education, and autonomous systems. In this talk, I will present a systematic, rigorous, and safety-centered exploration of machine unlearning in modern generative AI systems, with a primary focus on large language models (LLMs). Rather than treating unlearning as an isolated task, we position it as a multidisciplinary frontier shaped by the co-design of optimization, data, and model principles.
Date and Time
Location
Hosts
Registration
-
Add Event to Calendar