Descend or Rewind? Stochastic Gradient Descent Unlearning

This paper establishes (ε,δ)(\varepsilon, \delta) certified unlearning guarantees for stochastic versions of the Descent-to-Delete (D2D) and Rewind-to-Delete (R2D) algorithms across strongly convex, convex, and nonconvex loss functions by analyzing them as disturbed gradient systems and coupling their trajectories with retraining, ultimately revealing that D2D offers tighter guarantees for strongly convex functions while R2D is better suited for convex and nonconvex scenarios.

Siqiao Mu, Diego Klabjan

Published 2026-03-02
📖 5 min read🧠 Deep dive

Imagine you have a giant, super-smart student who has studied a massive library of books to become an expert. One day, a specific person (let's call him "Bob") realizes his book was used to train this student and demands, "I want my book removed from your memory! I have the right to be forgotten."

The student can't just delete one page from a million-page book and hope the rest makes sense. The old way to fix this was to make the student forget everything, throw away the library, and start studying from scratch without Bob's book. This is accurate, but it takes forever and costs a fortune in energy.

This paper introduces two clever shortcuts to "unlearn" Bob's book without restarting the whole school year. The authors compare two methods: "Descent-to-Delete" (D2D) and "Rewind-to-Delete" (R2D).

Here is the breakdown of their findings using simple analogies:

The Two Strategies

1. Descent-to-Delete (D2D): "The Backwards Walk"

  • The Idea: Imagine the student is standing at the very top of a mountain (the final trained model). To remove Bob's influence, the student tries to walk backwards down the slope, taking small steps to adjust their position.
  • The Problem: On a smooth, simple hill (a "convex" function), this works great. But on a real-world mountain with jagged peaks, hidden valleys, and tricky terrain (a "nonconvex" function, like deep neural networks), walking backwards is dangerous. The student might get stuck in a small hole (a local minimum) or wander off into a completely different valley. They end up in a spot that looks different from where they started, but it's not the right spot for the "Bob-less" version.
  • The Paper's Verdict: This method is risky for complex, modern AI models. It often fails to actually "forget" correctly and might just make the model worse at other tasks.

2. Rewind-to-Delete (R2D): "The Time Machine"

  • The Idea: Instead of walking backwards from the finish line, imagine the student had a time machine. They go back to a checkpoint from earlier in their studies (say, halfway through the semester). From that earlier point, they re-study the library without Bob's book.
  • Why it works: Because they are starting from a point where the student's knowledge was still flexible and not yet "stuck" in a specific pattern, they can smoothly integrate the new reality (no Bob) without getting lost. They are essentially re-doing the last part of the training, but starting from a safer, more stable place.
  • The Paper's Verdict: This is the winner for complex, modern AI. It is more reliable, computationally cheaper, and actually achieves the goal of "forgetting" without breaking the model.

The "Stochastic" Twist (The Noise Factor)

In the real world, students don't read books in perfect order; they read random pages (this is called Stochastic Gradient Descent or SGD). This randomness makes the math much harder.

  • The Challenge: The authors had to prove that even with this randomness, the "Rewind" method still guarantees that the final model is indistinguishable from one that never saw Bob's book at all.
  • The Secret Sauce: They used a mathematical trick called "Coupling." Imagine two students: one studying with Bob's book, and one studying without it. The authors proved that if you make them read the exact same random pages at the same time (except for Bob's book), their paths will stay very close together.
    • For simple hills, the "Backwards Walk" (D2D) is actually slightly more precise.
    • For jagged mountains (real-world AI), the "Time Machine" (R2D) is the only one that doesn't get lost.

The "Privacy Noise"

To make the unlearning mathematically "certified" (proving to a judge that Bob is truly forgotten), the algorithms add a tiny bit of "static" or "noise" to the final answer.

  • Think of this like blurring a photo just enough so you can't identify a specific face, but the photo still looks like a face.
  • The paper proves that the "Rewind" method needs less of this blurring (noise) to be safe, meaning the final model stays smarter and more accurate.

The Bottom Line

If you want to remove data from a simple, predictable model, you can try walking backwards (D2D). But if you are dealing with modern, complex AI (like the ones powering chatbots or image generators), you should Rewind to an earlier checkpoint and retrain from there (R2D).

Why does this matter?

  • Privacy: It helps companies comply with laws like GDPR (the "Right to be Forgotten") without spending millions of dollars retraining models from scratch.
  • Efficiency: It saves massive amounts of energy and time.
  • Safety: It ensures that when a user asks to be forgotten, they actually are, rather than the model just pretending to forget while secretly keeping the data in a hidden corner.

In short: Don't try to walk backwards off a cliff; use a time machine to go back to solid ground.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →