Zero-Shot Class Unlearning in CLIP with Synthetic Samples

Research output: Contribution to conferencePaperpeer-review

2 Downloads (Pure)

Abstract

Machine unlearning is a crucial area of research. It is driven by the need to remove sensitive information from models to safeguard individuals’ right to be forgotten under rigorous regulations such as GDPR. In this work, we focus on unlearning within CLIP, a dual vision-language encoder model trained on a massive dataset of image-text pairs using contrastive loss. To achieve forgetting we expand the application of Lipschitz regularization to the multimodal context of CLIP. Specifically, we smooth both visual and textual embeddings associated with the class intended to be forgotten relative to the perturbation introduced to the samples from that class. Additionally, importantly, we remove the necessity for real forgetting data by generating synthetic samples via gradient ascent maximizing the target class. Our forgetting procedure is iterative, where we track accuracy on a synthetic forget set and stop when accuracy falls below a chosen threshold. We employ a selective layers update strategy based on their average absolute gradient value to mitigate over-forgetting. We validate our approach on several standard datasets and provide thorough ablation analysis and omparisons with previous work.
Original languageEnglish
Number of pages28
Publication statusPublished - 11 Dec 2024
EventIEEE/CVF Winter Conference on Applications of Computer Vision - JW Marriott Starpass, Tucson, USA United States
Duration: 28 Feb 20254 Mar 2025
https://wacv2025.thecvf.com/

Conference

ConferenceIEEE/CVF Winter Conference on Applications of Computer Vision
Abbreviated titleWACV
Country/TerritoryUSA United States
CityTucson
Period28/02/254/03/25
Internet address

Keywords

  • CLIP
  • unlearning
  • zero-shot

Fingerprint

Dive into the research topics of 'Zero-Shot Class Unlearning in CLIP with Synthetic Samples'. Together they form a unique fingerprint.

Cite this