The Stability of Online Algorithms in Performative Prediction

This paper establishes an unconditional reduction proving that any no-regret online algorithm deployed in performative prediction settings converges to a performatively stable equilibrium, thereby demonstrating that such algorithms naturally prevent runaway feedback loops without requiring restrictive assumptions on how models influence data distributions.

Gabriele Farina, Juan Carlos Perdomo

Published 2026-03-02
📖 5 min read🧠 Deep dive

The Big Picture: The "Self-Fulfilling Prophecy" Problem

Imagine you are a weather forecaster.

  • Scenario A (Normal): You predict rain. People bring umbrellas. It rains. You were right. The world didn't change because of your prediction.
  • Scenario B (Performative): You predict a massive traffic jam. Because of your prediction, everyone decides to stay home or take a different route. Suddenly, the roads are empty! Your prediction caused the traffic jam not to happen.

This is the core problem the paper addresses: When algorithms make predictions, those predictions often change human behavior, which changes the data the algorithm sees next.

If a bank's AI predicts you are a "high risk" for a loan, you might be denied credit. Without credit, you can't build a good financial history, so you become high risk. The AI's prediction created the very reality it feared. This creates a feedback loop that can spiral out of control, making the system unstable and unpredictable.

The Old Way: Trying to Control the World

For years, researchers tried to fix this by assuming the world is "gentle." They assumed that if an algorithm changes its mind slightly, the world only changes slightly in response. They called this the "Lipschitz condition" (a fancy math way of saying "no sudden jumps").

The Analogy: Imagine trying to balance a broom on your hand.

  • The Old Assumption: The broom is made of soft rubber. If you tilt it a little, it wobbles a little, and you can easily correct it.
  • The Reality: In the real world (like medicine, education, or finance), the "broom" is made of glass. If you tilt it even a tiny bit, it might shatter or snap back violently.
  • The Problem: Recent research showed that if the world reacts violently (discontinuously), it is mathematically impossible to find a single, perfect model that stays stable. It's like trying to find a spot to balance a broom that keeps changing its shape.

The New Discovery: The "Chameleon Strategy"

The authors of this paper (Gabriele Farina and Juan Carlos Perdomo) found a brilliant workaround. They realized you don't need to find one perfect model that never changes. Instead, you should use a mixture (a random blend) of many different models.

The Analogy: The "Blind Taste Test"
Imagine a restaurant trying to find the perfect soup recipe.

  • The Old Way: The chef picks one recipe, serves it, sees how people react, and tries to tweak that one recipe forever. If the customers' tastes change wildly, the chef gets confused and keeps changing the recipe, never settling.
  • The New Way: The chef decides to serve a random mix of 100 different recipes every day. Some days it's spicy, some days it's mild.
    • The customers' reactions change based on the soup they get.
    • But because the chef is averaging over all the recipes, the overall feedback loop stabilizes. No single recipe is "wrong" because the system is designed to handle the chaos by spreading the risk.

The Magic Ingredient: "No-Regret" Algorithms

How do we create this mix? The paper uses a concept from online learning called "No-Regret."

Think of a gambler at a casino.

  • A "No-Regret" gambler doesn't try to predict the future perfectly. Instead, they just make sure that over time, they didn't miss out on a better strategy they could have used.
  • If they played the game 1,000 times, their total winnings are almost as good as if they had known the best move for every single hand in hindsight.

The paper proves a magical connection: If you use a "No-Regret" algorithm to update your models over time, and then you take a random sample from all the models you've ever created, that mix is guaranteed to be stable.

It doesn't matter if the world is chaotic, if the data jumps around, or if the rules change suddenly. As long as the algorithm is "smart enough" to minimize its regret over time, the resulting mix of models will stop the runaway feedback loop.

Why This Matters (The "Aha!" Moment)

  1. It Works Everywhere: The old methods failed if the data was "jumpy" (like a student passing or failing a class based on a strict cutoff score). This new method works even with those jumpy, discontinuous rules.
  2. It Explains Why We Don't Crash: You might wonder, "Why don't our current AI systems (like recommendation engines) crash the economy or society?" The paper suggests it's because these systems naturally act like "No-Regret" algorithms. They constantly tweak and retrain, and even though they are chasing a moving target, the average of their behavior naturally settles into a stable state.
  3. No Magic Assumptions Needed: We don't need to assume the world is nice and smooth. We just need the algorithm to keep learning and adapting.

Summary in One Sentence

Instead of trying to find one perfect, unchanging crystal ball that predicts the future (which is impossible when the future changes because of the prediction), we should use a "No-Regret" learning process that constantly adapts; the average of all the models it creates along the way will naturally stabilize the system, preventing runaway feedback loops even in chaotic environments.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →