Confidence intervals for the Poisson distribution

This paper addresses confusion among physicists regarding Poisson sampling results by comparing various techniques and recommending Garwood's confidence intervals as the most consistent and intuitive method for summarizing data.

Original authors: Frank C. Porter

Published 2026-04-22
📖 6 min read🧠 Deep dive

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

Imagine you are a detective trying to solve a mystery: How many times did a rare event happen?

In the world of physics (and many other sciences), events like radioactive decays or particle collisions happen randomly. Sometimes you see them 10 times, sometimes 2 times, and sometimes 0 times. This randomness follows a specific rule called the Poisson distribution.

The problem isn't counting the events; it's telling the story of what that count means. If you counted 3 events, does that mean the "true" rate of the universe is exactly 3? Probably not. It could be 2, or 4, or even 10, just because of random luck.

Scientists need a way to say: "Based on seeing 3 events, the true rate is likely somewhere between X and Y." This range is called a Confidence Interval.

For decades, physicists have been arguing about the best way to draw this range. Some methods are too wide (wasting data), some are too narrow (lying about precision), and some behave strangely when you look at them closely.

Frank Porter's paper is like a judge settling a courtroom dispute between all these different methods. He asks: "Which method tells the truth about the measurement without getting confused by what we think the truth should be?"

Here is the breakdown of his verdict, using simple analogies.


1. The Core Conflict: Description vs. Interpretation

Imagine you are describing a photo of a blurry face.

  • Description: "The photo shows a blurry blob that looks like it might be a dog." (This is what the paper focuses on: describing the data exactly as it is).
  • Interpretation: "I am 95% sure that the blurry blob is actually a dog." (This is what people often want to do, but it requires guessing about the "truth" before seeing the data).

The author argues that most confusion comes from mixing these up. We should first describe the measurement objectively (the blurry blob) before trying to guess the truth (the dog). If we try to force the description to fit our physical beliefs (e.g., "The rate can't be negative!"), we end up with confusing, broken math.

2. The Contenders (The Methods)

The paper reviews many different "rulers" scientists use to measure the uncertainty. Here are the main characters:

  • The "Garwood" Ruler (The Old Reliable):

    • How it works: It's a classic, conservative method. It draws a wide net to make sure it never misses the true value.
    • Pros: It's consistent. If you look at the same data with different levels of certainty, the ranges fit inside each other perfectly (like Russian nesting dolls). The math behaves smoothly.
    • Cons: It's sometimes a bit too wide (over-covering), meaning it might say the answer is between 1 and 10 when it's really between 3 and 4. It's "safe" but not "tight."
  • The "Crow & Gardner" Ruler (The Tightrope Walker):

    • How it works: It tries to make the net as small as possible to be more precise.
    • Pros: It's often shorter (more precise) than Garwood.
    • Cons: It's chaotic. If you change the confidence level slightly (from 90% to 95%), the range might jump wildly or even exclude the most likely answer. It's like a tightrope walker who falls off if the wind blows a little.
  • The "Feldman-Cousins" Ruler (The Physical Enforcer):

    • How it works: It forces the answer to stay in the "physical" zone (e.g., it won't allow a negative number of particles).
    • Pros: It feels intuitive to physicists who hate negative numbers.
    • Cons: When the data is weird (like seeing fewer events than the background noise), this ruler shrinks the range to almost zero. It tricks you into thinking you have super-precise knowledge when you actually have very little. It hides the fact that the background noise fluctuated wildly.
  • The "Bayesian" Ruler (The Believer):

    • How it works: It starts with a guess (a "prior") about what the answer might be, then updates it with data.
    • Pros: Great for making decisions.
    • Cons: It depends on your initial guess. If two people have different beliefs, they get different answers. The paper argues this is about belief, not description of the measurement.

3. The "Averaging" Trap

The paper also warns about a common mistake: Averaging results.
Imagine you have 10 different experiments, each with its own confidence interval. You might think, "I'll just average the middle numbers and the widths."

  • The Trap: If you do this with Poisson data, you can accidentally create a result that is less accurate than the individual parts. It's like averaging a bunch of blurry photos and expecting a sharp image. The math breaks down unless you go back to the raw data (the original counts) and re-calculate everything together.

4. The Verdict: Why Garwood Wins

After testing all these rulers against a list of "Desirable Properties" (like: Does it behave smoothly? Does it nest? Does it give sensible "p-values" which are like probability scores?), the author declares a winner:

The Garwood Interval.

Why?

  1. It's Honest: It describes the measurement without trying to force it into a "physical" box that distorts the math.
  2. It's Stable: If you tweak the confidence level, the answer changes smoothly. It doesn't jump around.
  3. It's Consistent: The ranges nest perfectly (a 90% range is always inside a 95% range).
  4. It Makes Sense: The "p-values" (probability scores) it generates are intuitive and continuous.

The Trade-off:
Yes, the Garwood interval is sometimes a little wider than necessary. But the author argues that it is better to be slightly too safe and consistent than to be precise but chaotic. A method that jumps around or gives weird answers when you look at it from a slightly different angle is dangerous for science.

Summary in One Sentence

When counting rare, random events, don't try to force the math to fit your physical beliefs; instead, use the Garwood method because it provides a stable, consistent, and honest description of the data, even if it's a little bit wider than the other options.

The Takeaway for Everyday Life:
When you are unsure about something, it's better to have a wide, reliable estimate that doesn't change when you look at it from a different angle, than a narrow, precise estimate that falls apart the moment you test it.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →