Spread them Apart: Towards Robust Watermarking of Generated Content

This paper proposes a robust, inference-time watermarking method for generative models that embeds detectable and user-identifiable marks into generated images without retraining, offering provable resistance to additive perturbations and synthetic removal attacks.

Mikhail Pautov, Danil Ivanov, Andrey V. Galichin, Oleg Rogov, Ivan Oseledets

Published 2026-03-02
📖 5 min read🧠 Deep dive

Imagine you just bought a high-end 3D printer that can create incredibly realistic paintings, sculptures, and photos. It's so good that you can't tell the difference between a human artist's work and what your machine spits out.

Now, imagine a problem: A dishonest person uses your machine to print a fake masterpiece, claims they painted it themselves, and sells it. Or, someone uses it to create a "deepfake" of a celebrity saying something they never said. How do you prove who actually made it?

This paper introduces a solution called "Spread them Apart." Think of it as a clever, invisible security tag that gets baked into the image while it is being created, rather than stamped on top afterward.

Here is the breakdown using simple analogies:

1. The Problem: The "Perfect Forgery"

Generative AI (like the ones that make images from text) has gotten so good that fake images look real.

  • The Risk: Without proof, anyone can take an AI image, claim it's theirs, and break copyright laws. Or, they could try to scrub away any hidden proof that it was AI-made.

2. The Solution: The "Invisible DNA"

The authors propose a method to embed a digital watermark directly into the image's "DNA" during the creation process.

  • How it works: When you ask the AI to make a picture, the system doesn't just make the picture; it secretly tweaks the internal math of the image to fit a specific pattern unique to you (the user).
  • The Analogy: Imagine a baker making a cake. Instead of just baking it, they arrange the sprinkles inside the cake in a specific, secret pattern that only they know how to read. If someone tries to eat the cake (or cut it up), the pattern is still there inside.

3. The Secret Sauce: "Spread Them Apart"

This is the core trick of the paper.

  • The Old Way: Usually, watermarks are like a faint signature on the surface of a painting. If you wash the painting or change the lighting, the signature might fade.
  • The New Way: The authors use a strategy called "Spread them Apart."
    • Imagine you have two specific pixels (tiny dots of color) in the image. Let's call them Pixel A and Pixel B.
    • To hide a "1" in your secret code, the AI makes sure Pixel A is slightly brighter than Pixel B.
    • To hide a "0", it makes Pixel B slightly brighter than Pixel A.
    • The Magic: The AI ensures the difference between them is big enough that even if someone tries to blur the image, change the brightness, or add noise, Pixel A will still be brighter than Pixel B. The relationship is "spread apart" enough to survive the attack.

4. Why It's Hard to Remove (The "Unbreakable Seal")

The paper proves mathematically that this method is robust against common tricks people use to hide watermarks:

  • Brightness/Contrast: If someone turns up the brightness, both pixels get brighter, but the difference between them stays the same.
  • Flipping Colors: If someone turns the image into a negative (black becomes white), the relationship flips, but the system knows to look for the flip.
  • Adversarial Attacks: Even if a hacker tries to use a super-computer to specifically target and erase this pattern, the math shows it's incredibly difficult to do without ruining the image itself.

5. The "Three-Way" Backup Plan

To make it even stronger, the paper suggests a "belt and suspenders" approach.

  • Instead of just hiding the secret in the pixels, they also hide it in the mathematical "shape" of the image (using frequency patterns).
  • The Analogy: It's like writing a secret message in three places:
    1. On the surface of a rock.
    2. Inside the rock's crystal structure.
    3. In the shadow the rock casts.
    • Even if someone chips off the surface (pixel attack) or melts the rock (geometric attack), the message might still be recoverable from the shadow or the crystal structure.

6. The Result: Who Made It?

When a suspicious image pops up, the owner of the AI system can run a "decoder":

  1. They look at the secret pairs of pixels.
  2. They check the pattern (Is A brighter than B? Or B brighter than A?).
  3. They reconstruct the secret code.
  4. The Verdict: If the code matches User #42, they know User #42 generated it. If the code is gibberish, they know it wasn't made by their system.

Summary

"Spread them Apart" is a way to bake a permanent, unremovable ID card into AI-generated images. It doesn't require retraining the AI; it just tweaks the final steps of creation to ensure that a secret relationship between pixels is strong enough to survive almost any attempt to scrub it away. It's the difference between a sticker on a car (easy to peel off) and the car's VIN number stamped into the frame (hard to remove without destroying the car).