Here is an explanation of the paper "Quantifying Uncertainty in AI Visibility," translated into simple, everyday language with some creative analogies.
The Big Idea: AI is a Fickle Friend, Not a Reliable Librarian
Imagine you ask a very smart, but slightly unpredictable friend (let's call him "AI") to recommend the best running shoes. You ask him the exact same question three times in a row.
- Ask 1: He says, "Go to Nike.com and Adidas.com."
- Ask 2: He says, "Check out Adidas.com and NewBalance.com."
- Ask 3: He says, "You should look at Nike.com and Brooks.com."
If you only asked him once, you might conclude, "Okay, Nike is the top recommendation!" But if you asked him ten times, you'd realize his answers are all over the place.
This is exactly what this paper is about.
For years, marketers and brands have treated AI search engines (like Google Gemini, Perplexity, and OpenAI's SearchGPT) like a static list. They ask, "How often does my website appear in the AI's answer?" and treat the answer as a fixed fact (e.g., "We have 12% visibility").
The paper argues this is a dangerous mistake. Because AI is "non-deterministic" (it uses randomness to generate answers), its citations change every time you ask. A single measurement is like taking one photo of a moving car; it doesn't tell you where the car is actually going.
The Core Problem: The "Snapshot" Trap
The authors say that when companies measure their "AI Visibility," they are usually taking a snapshot. They run 200 questions, count the hits, and say, "We are #1!"
But because the AI is flipping a coin behind the scenes to decide which sources to pick, that "12%" number is actually just a guess with a huge margin of error.
The Analogy: The Weather Forecast
Imagine you want to know if it's going to rain.
- The Old Way (Single Run): You look out the window for 5 minutes. It's sunny. You conclude, "It will be sunny all day."
- The New Way (This Paper): You realize the weather is chaotic. You need to look out the window for 5 minutes, then again 10 minutes later, then again 15 minutes later. You might see a cloud, then sun, then a drizzle. You realize, "It's actually a bit of a toss-up."
The paper shows that for AI search, the "weather" changes so fast that a single 5-minute look (a single run of 200 queries) is useless for making business decisions.
The Three Main Findings (Simplified)
1. The "Dice Roll" Effect (System Stochasticity)
The AI isn't broken; it's just designed to be random. Even if you ask the exact same question to the exact same AI at the exact same time, it might give you a different list of websites.
- The Metaphor: Think of the AI as a DJ playing a playlist. If you ask for "Jazz," the DJ might play Song A, Song B, and Song C. If you ask again, they might play Song B, Song D, and Song E. The "Jazz" vibe is there, but the specific songs change.
- The Result: A brand might appear in the first playlist but not the second. If you only listen to the first playlist, you think you're famous. If you listen to ten, you realize you're only famous half the time.
2. The "Confidence Interval" (The Safety Net)
The paper suggests we stop reporting single numbers (like "12% visibility") and start reporting ranges (like "12% visibility, plus or minus 4%").
- The Analogy: Imagine a dartboard.
- Old Way: You throw one dart. It hits the bullseye. You say, "I am a perfect aim!"
- New Way: You throw 100 darts. They are scattered in a circle around the bullseye. You say, "My aim is good, but I usually land within this circle."
- Why it matters: The paper found that for many brands, the "circle" is so big that Brand A (12%) and Brand B (9%) are actually in the same circle. Statistically, they are tied. But without the "circle" (confidence interval), companies waste money trying to beat a rival they are actually equal to.
3. The "Power Law" (The Long Tail)
The distribution of who gets cited looks like a pyramid. A few big websites get cited a lot, and hundreds of small websites get cited a little.
- The Metaphor: Think of a concert. The headliner (the top brand) gets 50% of the applause. The opening act gets 10%. The rest of the band gets 1% each.
- The Twist: The paper found that the "headliners" are actually quite stable (they show up most of the time), but the "opening acts" and the "band members" are incredibly volatile. One day a small blog is mentioned; the next day, it's gone. This makes measuring the "middle" of the market very difficult.
The "Content Change" Check (Did the Website Change?)
You might think, "Maybe the websites themselves changed! Maybe Runner's World updated their article, so the AI stopped citing them."
The authors checked this. They took "fingerprints" (digital checksums) of the websites to see if the content actually changed.
- The Result: The websites didn't change. The content was stable.
- The Conclusion: The chaos is coming from the AI, not the websites. The AI is the one flipping the coin, not the websites changing their minds.
What Should You Do? (Practical Advice)
If you are a marketer or a business owner using these tools, the paper gives you three rules:
- Stop trusting single numbers. If someone tells you, "We have 15% visibility," ask, "What is the margin of error?" If they don't know, don't trust the number.
- Run more tests. To get a reliable answer, you can't just ask 200 questions once. You need to ask questions over several days or run the test multiple times to see the pattern.
- Accept the noise. Sometimes, you won't be able to tell if Brand A is better than Brand B. They might just be statistically tied. Don't panic and change your strategy based on a tiny, noisy difference.
The Bottom Line
Generative AI is a powerful tool, but it's a noisy one. Treating its output as a fixed fact is like trying to measure the ocean's depth with a single ruler drop. You need to take many measurements, calculate the average, and admit that there is always some uncertainty.
In short: The AI is playing a game of chance. If you want to win, you need to understand the odds, not just look at the result of one roll of the dice.