ArXiv TLDR

Efficient Unlearning through Maximizing Relearning Convergence Delay

🐦 Tweet
2604.09391

Khoa Tran, Simon S. Woo

cs.LGcs.CV

TLDR

This paper introduces a new metric, relearning convergence delay, and an unlearning framework to efficiently remove data influence from models.

Key contributions

  • Introduces 'relearning convergence delay' to assess unlearning by tracking changes in weight and prediction space.
  • Proposes 'Influence Eliminating Unlearning' framework to remove data influence while maintaining accuracy.
  • Achieves unlearning by degrading forgetting set performance, using weight decay, and injecting noise.
  • Provides theoretical guarantees and empirical evidence of strong retention and resistance to relearning.

Why it matters

Current unlearning methods lack comprehensive evaluation. This paper offers a novel metric and framework that provide a deeper understanding of unlearning, crucial for robust data privacy and model security. It ensures forgotten data cannot be easily recovered.

Original Abstract

Machine unlearning poses challenges in removing mislabeled, contaminated, or problematic data from a pretrained model. Current unlearning approaches and evaluation metrics are solely focused on model predictions, which limits insight into the model's true underlying data characteristics. To address this issue, we introduce a new metric called relearning convergence delay, which captures both changes in weight space and prediction space, providing a more comprehensive assessment of the model's understanding of the forgotten dataset. This metric can be used to assess the risk of forgotten data being recovered from the unlearned model. Based on this, we propose the Influence Eliminating Unlearning framework, which removes the influence of the forgetting set by degrading its performance and incorporates weight decay and injecting noise into the model's weights, while maintaining accuracy on the retaining set. Extensive experiments show that our method outperforms existing metrics and our proposed relearning convergence delay metric, approaching ideal unlearning performance. We provide theoretical guarantees, including exponential convergence and upper bounds, as well as empirical evidence of strong retention and resistance to relearning in both classification and generative unlearning tasks.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.