Certifying the Right to Be Forgotten: Primal-Dual Optimization for Sample and Label Unlearning in Vertical Federated Learning

This paper proposes FedORA, a primal-dual optimization framework that enables efficient and theoretically certified sample and label unlearning in Vertical Federated Learning by introducing a novel uncertainty-promoting loss function and adaptive strategies to minimize computational overhead while preserving model utility.

Yu Jiang, Xindi Tong, Ziyao Liu, Xiaoxi Zhang, Kwok-Yan Lam, Chee Wei Tan

Published Tue, 10 Ma
📖 5 min read🧠 Deep dive

Imagine you are part of a massive, collaborative cooking club. Everyone brings a different ingredient to the table to create a giant, delicious stew (the AI model).

  • Party A brings the spices.
  • Party B brings the vegetables.
  • Party C (the chef) brings the meat and the final seasoning.

Together, you make a stew that tastes amazing. But then, someone says, "Hey, I don't want my specific batch of carrots in that stew anymore. Please remove them." Or maybe, "I never want any carrots in this recipe again."

This is the "Right to be Forgotten." In the world of Artificial Intelligence, if you want your data removed, the model shouldn't just ignore your data; it should act as if it never knew you existed.

The Problem: The "Re-cook" Dilemma

In a normal kitchen, if you want to remove an ingredient, you have two bad options:

  1. The "Start Over" Method: Throw away the whole pot and start cooking from scratch with the remaining ingredients. This is perfect (the stew tastes exactly right without the carrots), but it takes forever and wastes a ton of energy.
  2. The "Scrape It Out" Method: Try to fish the carrots out of the finished stew. This is fast, but you might accidentally rip out the meat or leave chunks of carrot behind. The stew might taste weird or be unsafe.

In Vertical Federated Learning (VFL), this is even harder because the ingredients are split up. You can't just reach into the pot; you have to coordinate with everyone holding a different part of the recipe. Existing methods often try to "scrape" the data out by aggressively pushing the model to forget, which often breaks the model (making it forget too much or become unstable).

The Solution: FedORA (The Smart Chef)

The authors propose a new method called FedORA. Think of FedORA not as a chef trying to fish out carrots, but as a smart, mathematical recipe adjustment system.

Here is how FedORA works, using simple analogies:

1. The "Confusion" Strategy (Primal-Dual Optimization)

Most old methods try to make the model hate the data it needs to forget. They tell the model, "If you see a carrot, scream 'WRONG!'" This is like trying to unlearn a song by playing it backwards at full volume. It often causes the model to get confused about everything else, ruining the stew.

FedORA's approach is different: It tells the model, "If you see a carrot, just be confused."

  • Instead of forcing a wrong answer, FedORA encourages the model to say, "I have no idea what this is."
  • The Analogy: Imagine a student taking a test. Instead of forcing them to write the wrong answer for a specific question (which might mess up their confidence on other questions), you tell them to leave it blank or guess randomly. The goal is uncertainty, not error. This removes the "memory" of the carrot without shaking the foundation of the whole recipe.

2. The "Tension" Meter (Dual Variables)

FedORA uses a mathematical "tension meter."

  • If the model is still remembering the carrot too well, the tension meter goes up, and the system applies more pressure to make it forget.
  • If the model is already confused enough, the tension meter goes down, and the system stops pushing.
  • The Analogy: It's like a thermostat. If the room is too hot (remembering too much), the AC turns on. If it's cool enough, the AC turns off. This ensures the model forgets just enough without freezing the whole house (ruining the model's performance on the other ingredients).

3. The "Smart Batch" Cooking (Asymmetric Design)

When you are cooking a huge pot of stew, you don't need to taste-test every single drop of the remaining soup to make sure it still tastes good. You only need to taste a few spoonfuls.

  • FedORA's Trick: It processes the "to-be-removed" carrots (the bad data) with full attention, checking every single one. But for the rest of the good ingredients, it only samples a small portion to check the flavor.
  • The Result: This saves a massive amount of time and energy (computing power) because the system isn't wasting resources re-tasting the whole pot.

Why This Matters

  • It's Fast: It doesn't require starting over from scratch.
  • It's Safe: It doesn't break the model's ability to recognize other things (like potatoes or beef).
  • It's Certified: The math proves that the result is almost as good as if you had started from scratch, but much faster.

The Bottom Line

FedORA is like a magical, efficient kitchen assistant that can remove specific ingredients from a complex, collaborative recipe without ruining the taste of the final dish. It does this by teaching the model to be politely confused about the unwanted data, rather than aggressively fighting it, and by only checking the "good" parts of the recipe just enough to keep things running smoothly.

This ensures that in our digital world, when you ask to be forgotten, the AI actually forgets you, without forgetting how to be helpful to everyone else.