The field of machine learning is moving towards developing methods for efficient and effective removal of unwanted knowledge from trained models, also known as machine unlearning. This is driven by the need to comply with data privacy regulations and protect sensitive information. Recent research has focused on developing innovative approaches to unlearning, including methods that can remove specific data from trained models without requiring access to the original training dataset.
Noteworthy papers in this area include: Assessing Representation Stability for Transformer Models, which introduces a model-agnostic detection framework for identifying adversarial examples. Unlearning at Scale, which presents a reproducible systems approach to implementing the right to be forgotten in large language models. Unlearning Comparator, which provides a visual analytics system for comparative evaluation of machine unlearning methods. CRISP, which introduces a parameter-efficient method for persistent concept unlearning using sparse autoencoders. Efficient Knowledge Graph Unlearning with Zeroth-order Information, which presents an efficient knowledge graph unlearning algorithm using Taylor expansion and zeroth-order optimization. Towards Source-Free Machine Unlearning, which presents a method for source-free unlearning that can estimate the Hessian of the unknown remaining training data.