Advances in Machine Unlearning

The field of machine learning is moving towards developing methods for efficient and effective removal of unwanted knowledge from trained models, also known as machine unlearning. This is driven by the need to comply with data privacy regulations and protect sensitive information. Recent research has focused on developing innovative approaches to unlearning, including methods that can remove specific data from trained models without requiring access to the original training dataset.

Noteworthy papers in this area include: Assessing Representation Stability for Transformer Models, which introduces a model-agnostic detection framework for identifying adversarial examples. Unlearning at Scale, which presents a reproducible systems approach to implementing the right to be forgotten in large language models. Unlearning Comparator, which provides a visual analytics system for comparative evaluation of machine unlearning methods. CRISP, which introduces a parameter-efficient method for persistent concept unlearning using sparse autoencoders. Efficient Knowledge Graph Unlearning with Zeroth-order Information, which presents an efficient knowledge graph unlearning algorithm using Taylor expansion and zeroth-order optimization. Towards Source-Free Machine Unlearning, which presents a method for source-free unlearning that can estimate the Hessian of the unknown remaining training data.

Sources

Assessing Representation Stability for Transformer Models

Unlearning at Scale: Implementing the Right to be Forgotten in Large Language Models

Unlearning Comparator: A Visual Analytics System for Comparative Evaluation of Machine Unlearning Methods

CRISP: Persistent Concept Unlearning via Sparse Autoencoders

Efficient Knowledge Graph Unlearning with Zeroth-order Information

Towards Source-Free Machine Unlearning

Built with on top of