Regression to the Mean: The Statistical Force You Keep Mistaking for Progress
C. PearsonYour worst employee had a great quarter. You praised them, told the team, maybe gave a small bonus. Next quarter? Back to underperforming. You concluded the praise worked — temporarily — but that sustained motivation is hard.
Photo by Markus Winkler on Pexels.
That conclusion is almost certainly wrong.
What you witnessed wasn't a motivational arc. It was regression to the mean, one of the most pervasive and least-discussed forces in statistics. It doesn't care about your interventions. It doesn't care about your praise or your punishment. It operates quietly in the background of almost every dataset you will ever touch, and it has been making humans draw false conclusions for centuries.
What It Actually Is
When an observation sits far from the average — unusually high, unusually low — a large portion of that distance is often explained by random variation, not signal. Measure the same thing again, and that random component is likely to be smaller. The value drifts back toward the mean. Not because anything changed. Because extreme values are, by definition, partly a product of luck.
Francis Galton noticed this in 1886 while studying the heights of parents and children. Tall parents tended to have children shorter than themselves. Short parents tended to have taller children. He called it "regression toward mediocrity." The name softened over time; the phenomenon did not.
Where It's Destroying Your Analysis Right Now
Think about how many real decisions hinge on measuring performance at its worst moment — and then treating whatever happens next as evidence of an intervention working.
A patient with dangerously high blood pressure comes in for treatment. You prescribe something. Their pressure drops. Was it the drug, or were they simply at a temporary peak when they walked through your door? Clinical trials exist precisely to answer this, yet observational studies keep missing it.
Coaches punish athletes for bad performances and praise good ones. The punished athletes improve; the praised ones regress. The coach walks away convinced that punishment works better than reward. Daniel Kahneman spent years unpacking this exact illusion — he watched it happen in real time with Israeli Air Force instructors and called it one of the most important insights of his career.
Schools identify "failing" students using a single bad test score, enroll them in intensive programs, then measure improvement. The improvement is real. The attribution is murky at best.
graph TD
A[Extreme Observation] --> B{Is it signal or noise?}
B --> C[Mostly Signal]
B --> D[Mostly Noise]
C --> E[Intervention may have real effect]
D --> F[Value regresses toward mean regardless]
F --> G[False attribution of cause]
The Trap Is Structural, Not Stupid
Here's what makes this so corrosive: the trap isn't a failure of intelligence. It's a failure of study design. You intervene on extreme cases — because of course you do, those are the ones that need attention — and extreme cases regress. The timing of your intervention is perfectly calibrated to make it look like it worked.
This is why randomized controlled trials randomize. Not just to balance confounders, but to break the link between "extreme performance triggered the intervention" and "regression made the performance improve."
Without a control group selected the same way, you cannot separate your effect from the gravitational pull of the mean.
How to Stop Falling For It
A few practical habits that help:
Measure twice before you act. If you're selecting people, schools, branches, or products based on a single extreme reading, take a second measurement before the intervention. Regression will already be visible — and you haven't done anything yet.
Ask what a do-nothing group would show. Even a rough historical baseline matters. If the bottom 10% of performers last year averaged a 15% improvement the following year without any program, your program's 18% improvement is less impressive than it looks.
Be skeptical of "turnaround" stories. The narrative of the struggling division that got a new manager and bounced back is everywhere in business writing. Some of those are real. Many are regression, repackaged as leadership.
The deeper issue is that we are wired to find causes. Random variation is an unsatisfying explanation — it doesn't suggest action, doesn't assign credit, doesn't make a good case study. So we overlay a story. The intervention worked. The coach's method was sound. The drug performed.
Sometimes those stories are true. Often they are regression wearing a disguise.
Next time something improves after you touch it, sit with the uncomfortable question: would it have improved anyway?
Get Mean Methods in your inbox
New posts delivered directly. No spam.
No spam. Unsubscribe anytime.