When correcting for regression to the mean is worse than no correction at all

This paper argues that common statistical corrections for regression to the mean are often flawed or impractical, proposing instead that researchers should evaluate uncorrected data against a structural null expectation derived from measurement repeatability to avoid systematic bias and inflated error rates.

Original authors: José F. Fontanari, Mauro Santos

Published 2026-03-03
📖 5 min read🧠 Deep dive
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

Imagine you are a coach trying to figure out if your training program works differently for different athletes. You measure their speed before training (x1x_1) and after training (x2x_2). You want to know: Do the slowest runners improve the most? (This is called "compensatory growth" or "catch-up").

To answer this, you calculate the "change" (how much faster they got) and see if it relates to their starting speed.

Here is the problem: Your stopwatch is slightly shaky. Sometimes you record a runner as slower than they really are just because of a bad start or a gust of wind. This is Measurement Error.

Because of this shaky stopwatch, a statistical illusion called Regression to the Mean (RTM) happens. It's like a magic trick where the data pretends to show a relationship that isn't really there. If you pick the slowest runners (who likely had a bad day or a bad measurement), they will naturally look faster the next time you measure them, even if you do nothing. They aren't actually improving; they are just "regressing" back to their average speed.

This paper argues that trying to "fix" this illusion with standard math tricks often makes things worse.

The Three Characters in This Story

The authors analyze three ways scientists try to solve this problem:

1. The "Crude" Slope (The Unadjusted View)

This is just looking at the raw data without doing anything fancy.

  • The Problem: Because of the shaky stopwatch, the data will almost always show a fake negative line. It will look like the slowest runners improved the most, even if they didn't.
  • The Analogy: Imagine looking at a reflection in a funhouse mirror. The image is distorted, but at least you know you're looking at a distorted image.

2. The "Berry et al." Method (The Popular Fix)

This is the method most ecologists and biologists currently use. It tries to mathematically "subtract" the illusion from the data.

  • The Paper's Verdict: This is dangerous.
  • The Analogy: Imagine you have a blurry photo of a cat. The Berry method is like using a filter that tries to sharpen the image, but it doesn't know how blurry the photo is.
    • If the photo is only slightly blurry, the filter might sharpen it too much, turning the cat into a tiger (creating a fake biological discovery).
    • If the photo is very blurry, the filter might erase the cat entirely, making it look like there was no animal there at all (hiding a real discovery).
  • The Result: This method is unreliable. It often creates "fake" scientific findings or hides real ones, leading to wrong conclusions about how animals grow or age.

3. The "Blomqvist" Method (The Perfect Fix)

This method is mathematically perfect. It can remove the illusion completely.

  • The Catch: To use it, you need to know exactly how shaky your stopwatch is (the measurement error).
  • The Analogy: This is like having a filter that can perfectly restore a blurry photo, but only if you know the exact model of the camera lens that caused the blur.
  • The Problem: In real life, we rarely know the exact "shakiness" of our measurements. Also, if you have a small group of athletes (a small sample size), this method gets very jittery and unstable. It's like trying to balance a pencil on its tip; it's theoretically possible, but in practice, it falls over easily.

The Authors' Solution: The "Reality Check"

Instead of trying to magically fix the data (which often introduces new errors), the authors suggest a smarter approach: The Reality Check.

Don't try to calculate the "perfect" answer. Instead, ask: "Is my result so strong that it couldn't possibly be explained just by my shaky stopwatch?"

  1. Estimate your "Repeatability": How consistent is your measurement? If you measure the same lizard's heat tolerance twice, do you get the same number? If the answer is "not very," your "Repeatability" is low.
  2. Calculate the "Fake" Slope: Based on how shaky your measurement is, calculate what the slope should look like if there were zero real biological effect. (The paper calls this the "structural null").
  3. Compare:
    • If your observed data looks just like the "Fake" slope, then stop. You haven't found a biological truth; you've just found the noise of your measurement.
    • If your data is way different from the "Fake" slope, then you might have found something real.

Real-World Examples from the Paper

  • Lizards: Scientists thought lizards with high heat tolerance couldn't get any better (a trade-off). The paper shows that this "trade-off" might just be the result of measurement noise. If the lizards' heat tolerance is hard to measure precisely, the data will look like they can't improve, even if they can.
  • Birds: Scientists thought birds with long telomeres (a marker of aging) lost them faster. The paper shows that when you account for the "shaky" measurement, the evidence for this rule disappears. It might just be statistical noise.

The Big Takeaway

"Correction" is not always better than "No Correction."

If you don't know how precise your measurements are, trying to "correct" for regression to the mean is like trying to fix a leaky roof by painting the ceiling. You might make the ceiling look nice, but the roof is still leaking, and now you've added paint to the mess.

The authors' advice:

  1. Stop blindly applying the "Berry" correction.
  2. Focus on Repeatability. Before you claim a biological discovery, you must know how reliable your measurements are.
  3. If you can't measure your error precisely, admit the uncertainty. Don't claim you found a "trade-off" or a "compensatory growth" if your data could easily be explained by a shaky ruler.

In short: Know your tools before you try to fix the picture.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →