Experimental evidence of progressive ChatGPT models self-convergence

This paper presents experimental evidence that recent ChatGPT models exhibit "model self-convergence," a phenomenon where their text outputs become increasingly similar and less diverse over time, likely due to the accumulation of synthetic data generated by LLMs in their training datasets.

Konstantinos F. Xylogiannopoulos, Petros Xanthopoulos, Panagiotis Karampelas, Georgios A. Bakamitsos

Published 2026-03-16
📖 5 min read🧠 Deep dive

The Big Idea: AI is Getting "Stuck in a Loop"

Imagine a group of chefs (the AI models) who are famous for writing recipes. For years, they learned by reading millions of cookbooks written by real humans. They were creative, varied, and could describe a "chocolate cake" in a thousand different delicious ways.

However, the internet has changed. Now, instead of just reading human cookbooks, these chefs are starting to read recipes written by other chefs who were trained on the internet.

This paper argues that we are seeing a phenomenon called "Model Self-Convergence." In plain English: The AI models are starting to sound more and more like each other, losing their ability to be creative or unique, because they are eating their own tail.

The Experiment: The "Paraphrase Test"

To prove this, the researchers set up a simple experiment:

  1. The Source Material: They took 443 summaries of classic books (like Bleak House) written by real humans. These are the "original recipes."
  2. The Chefs: They asked different versions of ChatGPT (from the old 2022 version to the newest 2025 versions) to rewrite these summaries.
  3. The Twist: They asked the AI to do this twice:
    • Temperature 0 (The Robot Mode): The AI tries to be as predictable and logical as possible.
    • Temperature 1 (The Creative Mode): The AI is told to be random, creative, and take risks.
  4. The Measurement: They used a special ruler (called "Similarity Percentage Ratio") to measure how similar the AI's rewrites were to each other.

The Findings: The "Echo Chamber" Effect

Here is what they found, using a simple analogy:

1. The Old AI (2022-2023): The Improvising Jazz Musician
When the older models were asked to be creative (Temperature 1), they were like jazz musicians. If you asked them to play a song, they would improvise. One time they might play it fast, another time slow, with different notes. Even if they were rewriting the same story, the results were very different from each other. They had diversity.

2. The New AI (2024-2025): The Broken Record
When the newest models were asked to be creative, they sounded like a broken record.

  • The Problem: Even when told to be random, the new models kept using the exact same phrases, sentence structures, and word choices.
  • The Result: If you asked the newest AI to rewrite a story five times, the five results were almost identical to each other. They had lost their "stochasticity" (their ability to be random).

3. The "Self-Convergence" Phenomenon
The paper calls this Model Self-Convergence.

  • Analogy: Imagine a room full of mirrors facing each other. If you stand in the middle, you see infinite reflections of yourself. The newer AI models are like those mirrors. They are trained on data that is increasingly filled with text they themselves generated.
  • Because the internet is now flooded with AI-written text (students using AI for homework, bloggers using AI for articles), the AI is learning from its own output. It's like a student who only studies the answers written by previous students, rather than the original textbook. Eventually, they all start giving the exact same answer.

Why Does This Matter?

The researchers found something scary: The newer models are getting worse at being creative, not better.

  • The "Gibberish" vs. "Blandness" Distinction: You might have heard of "Model Collapse," where AI starts spitting out total nonsense (gibberish). This paper says we aren't quite there yet. Instead, the AI is becoming bland. It's not making mistakes; it's just becoming a boring echo chamber.
  • The Loss of Innovation: If an AI can't generate diverse ideas, it can't be truly innovative. It will just recycle the same patterns it saw on the internet.
  • The Vicious Cycle: The more we use AI, the more AI text floods the internet. The more AI text floods the internet, the more future AIs get trained on it. The more they get trained on it, the less unique they become.

The Conclusion: A Warning for the Future

The paper concludes that unless we can find a way to train AI only on fresh, human-created data (which is getting harder to find), we risk a future where our digital assistants all sound exactly the same.

The Metaphor:
Think of the internet as a giant library.

  • Before: The library was full of books written by humans. The AI was a student reading all of them, learning to write in many different styles.
  • Now: The library is being filled with photocopies of the AI's own notes. The student is now reading only those photocopies.
  • The Result: The student stops learning new styles and starts repeating the same few sentences over and over, thinking they are the only words that exist.

This paper is a wake-up call: We need to keep the "human voice" in the training data, or the AI will eventually lose its voice entirely.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →