Critical Confabulation: Can LLMs Hallucinate for Social Good?

This paper proposes "critical confabulation," a framework where LLMs are guided to generate evidence-bound, speculative narratives to fill archival gaps regarding marginalized historical figures, demonstrating that controlled hallucinations can support ethical knowledge production without sacrificing historical fidelity.

Peiqi Sui, Eamon Duede, Hoyt Long, Richard Jean So

Published 2026-03-09
📖 5 min read🧠 Deep dive

Imagine you are trying to tell the story of a life, but someone has ripped out several pages from the middle of the biography. The pages are gone, and the history books are silent about what happened during those missing years. This is the reality for many "hidden figures" in history—people whose stories were erased by racism, poverty, or political violence.

This paper asks a bold question: Can we use Artificial Intelligence (AI) to fill in those missing pages in a way that is helpful, rather than just making things up?

Here is the breakdown of their idea, using simple analogies.

1. The Problem: The "Silent Archive"

Think of history as a giant library. For centuries, the librarians (historians and governments) only kept books about powerful people. The stories of enslaved people, the poor, and marginalized communities were often thrown in the trash or never written down.

When historians try to tell these stories today, they hit "blank spots." They know a person existed, but they don't know what they did on a specific Tuesday in 1854. Traditional history says, "If there is no evidence, we can't say anything." But the authors argue that this silence hurts us. It leaves the "hidden figures" invisible.

2. The Solution: "Critical Confabulation"

Usually, when AI makes things up, we call it a "hallucination," and we treat it like a bug. If an AI says the moon is made of cheese, that's a failure.

But the authors propose a new concept called Critical Confabulation.

  • The Analogy: Imagine you are a detective trying to solve a cold case. You have a witness who remembers the suspect wore a red hat and ran down the street, but they forgot why the suspect ran.
  • The Old Way: The detective says, "I don't know, so I won't guess."
  • The Critical Confabulation Way: The detective uses all the known facts (red hat, running, the time of day, the neighborhood) to construct a plausible story of why the suspect ran. They aren't claiming it's 100% proven fact; they are offering a "best guess" narrative to help visualize the missing piece.

The authors want to use AI to do this for history. They want the AI to look at the "gaps" in the records and write a story that fits the context, helping us imagine what life might have been like for those erased people.

3. The Experiment: The "Fill-in-the-Blank" Test

To see if AI is good at this, the researchers created a test.

  • The Setup: They took real, unpublished historical documents about Black intellectuals and activists (people the AI likely hasn't seen before).
  • The Game: They took a timeline of a person's life and erased one event, replacing it with a black box: [MASKED].
  • The Task: They asked various AI models to guess what happened in that black box.
  • The Goal: Did the AI just make up nonsense? Or did it write a story that felt true to the character and the time period, even if it couldn't be 100% proven?

4. The Results: AI Can Be a "Creative Historian"

The results were surprising and promising:

  • It's hard, but possible: The AI didn't get it right every time (only about 50-60% of the time), but it was often able to generate stories that felt "right" and matched the tone of the era.
  • Hints help: When the researchers told the AI, "This missing event was about a job change" or "This was about a family argument," the AI got much better at guessing.
  • The "Small" models are smart: Some smaller, open-source AI models performed just as well as the massive, expensive ones. This is great news because it means this tool could be accessible to more researchers.
  • No cheating: The researchers made sure the AI hadn't just memorized the answers from its training data. They used a "detective" method to ensure the AI was actually reasoning, not just reciting facts it already knew.

5. Why This Matters

The authors aren't saying AI should replace historians. Instead, they see AI as a co-pilot for storytelling.

  • For Historians: It's a tool to brainstorm ideas. If an AI suggests, "Maybe this person attended a secret meeting in 1920," a human historian can then go look for evidence to prove or disprove it. It turns a blank page into a hypothesis to investigate.
  • For Society: It helps "re-humanize" history. Instead of just seeing a list of dates and names, we get a narrative that helps us understand the lives of people who were systematically silenced.

The Bottom Line

Think of this paper as a new kind of archaeological brush. For a long time, we thought AI's tendency to "make things up" was a flaw. This paper argues that if we use that "making up" skill carefully—grounded in real facts and ethical boundaries—it can actually help us recover lost voices and fill in the gaps of our collective memory.

It's not about replacing the truth with fiction; it's about using imagination to find the truth that was hidden in the silence.