◆ Powerful

The Hot Hand Fallacy: And Why It Is Different

The belief that a player on a hot streak will keep performing well was declared a cognitive illusion in 1985. Then in 2018 a mathematician found a flaw in the original analysis. The hot hand may be real after all. This unit covers what changed, why it matters, and how it connects to a deeper point about dependent processes.

Time: 13 minutes
Requires: Unit 4.4

Opening Hook

In 1985, Thomas Gilovich, Robert Vallone, and Amos Tversky published one of the most celebrated findings in the psychology of judgment. They had analysed shooting records from the Philadelphia 76ers, surveyed basketball fans, and run controlled shooting experiments with Cornell varsity players. Their conclusion was unambiguous: the hot hand is an illusion.

Players who had just made several shots in a row were no more likely to make the next one. The streak you were watching was a pattern your brain had invented, the same clustering illusion that makes random sequences look meaningful. Belief in the hot hand was a cognitive bias, full stop. The paper was widely cited, widely taught, and widely treated as settled.

For thirty years, that was the story.

Then in 2018, Joshua Miller and Adam Sanjurjo published a paper in Econometrica, one of the most rigorous journals in economics, with an uncomfortable finding. The original analysis had a flaw. A subtle but real statistical bias had caused Gilovich and colleagues to systematically underestimate streakiness in their data. When that bias was corrected, the conclusion flipped. The hot hand showed up. There was, and perhaps always had been, significant evidence of streak shooting.

The original debunking had itself been partially debunked.

This is not a story about sloppy science. The bias Miller and Sanjurjo identified is genuinely non-obvious. The original researchers were careful and serious. What the episode illustrates is something more useful: that even well-designed analyses can contain subtle structural errors that only emerge later, and that the confidence we place in any single finding should always be provisional. The conclusion is not “believe in the hot hand.” It is “believe in findings held lightly, and understand what the argument actually rests on.”

The Concept

The hot hand effect is the claim that a person who has recently been performing well is more likely than usual to continue performing well on the next attempt. In basketball, the idea is that after three made shots in a row, a player’s probability of making the fourth is genuinely elevated. The word “hot” describes a temporary state of heightened performance. The claim is that this state is real and predictable.

Gilovich, Vallone, and Tversky (1985) set out to test this claim rigorously. They looked at the Philadelphia 76ers’ home games from the 1980—81 season and computed the probability that each player would hit a shot given that his previous one, two, or three shots were hits, and compared that to his hit rate after misses. If the hot hand were real, you would expect to see elevated success rates following streaks of hits. They found no such pattern. The probabilities were essentially the same regardless of recent history. They also found that 91 per cent of basketball fans believed the hot hand was real, and that the fans’ intuitions about how much a hot streak would boost performance were wildly inflated.

For three decades, this finding stood. The hot hand was taught in psychology and behavioural economics courses as a textbook example of the clustering illusion: the human tendency to see meaningful patterns in genuinely random sequences.

The problem Miller and Sanjurjo identified is called streak selection bias, and it is subtle enough to deserve a careful explanation.

Suppose you flip a fair coin 100 times and record the results as a sequence. Now you want to measure whether heads is more likely after a streak of three previous heads. To do this, you find every position in the sequence that was preceded by three heads, and you record what happened at that position. Then you calculate what fraction of those positions were also heads.

Here is the problem. In a finite sequence, there is a systematic mathematical reason why this fraction will, on average, come out lower than 50 per cent, even for a genuinely fair coin with no memory. The act of identifying “the flip after a streak of heads” creates a selection effect that biases the sample. The streak of heads that precedes your measurement point has, by its own existence, made it slightly more likely that a tail will immediately follow, simply because of how the positions overlap and interact within the finite sequence. The bias is not large in any individual sequence, but it is consistent, and in the kind of sample sizes used in the original basketball study, it is large enough to matter.

When Miller and Sanjurjo applied the correct adjusted analysis to the Gilovich data, they found that players were actually hitting at meaningfully higher rates after streaks of makes than after streaks of misses. The bias in the original methodology had been masking that signal. Their reanalysis also found evidence of streak shooting in further basketball data sets.

The current state of the evidence is: the hot hand is probably real, at least in some domains, and probably smaller in magnitude than fans believe. But the question of exactly how large and how consistent it is remains genuinely open.

Now for the distinction that gives this unit its title.

In Unit 4.4, you learned about the gambler’s fallacy: the erroneous belief that after a run of one outcome, the opposite outcome is “due.” The gambler’s fallacy treats independent events as if they are connected. A roulette wheel has no memory. A fair coin has no memory. Applying the logic of “due” to truly independent processes is always wrong.

The hot hand is a different structure entirely. A basketball player is not a roulette wheel. A shooter’s probability of making a basket depends on physical and mental states that do persist over short time windows: muscle memory, confidence, rhythm, fatigue, focus. These are not independent across attempts. The question “is the hot hand real?” is a question about whether those states produce autocorrelation in performance, meaning whether consecutive outcomes influence each other.

Autocorrelation means that observations close together in a sequence are correlated with each other. Daily temperatures are autocorrelated: if it was warm today, it is probably warm tomorrow, not because yesterday’s temperature caused tomorrow’s but because both are expressions of the same underlying weather system. A player’s shots within a single game may be autocorrelated because both draws on the same underlying physiological and psychological state.

The gambler’s fallacy is an error because it applies dependent-process logic to independent processes. The hot hand question is empirically legitimate because it asks whether a specific human performance process is actually dependent or not. These are different questions. Whether the hot hand turns out to be large, small, or absent in any particular domain, the question is a proper one, unlike the gambler’s question, which is based on a category error from the start.

The lesson is not “streaks are always meaningful.” It is “the meaningfulness of a streak depends entirely on whether the underlying process is independent or not, and that is always a question that needs to be investigated, not assumed.”

Why It Matters

The most consequential application of hot hand thinking outside sport is fund manager performance.

When an investment fund outperforms the market for several years running, the marketing materials write themselves. The manager has a gift. She sees what others miss. Past performance, the small print hurriedly notes, is not necessarily indicative of future results. But that is not quite how it is presented.

The structure of the claim is identical to the hot hand argument. Past success predicts future success. The question is whether fund management is more like basketball or more like a roulette wheel.

The evidence is sobering. Large-scale studies of fund performance, including work by SPIVA and the research behind the S&P Indices vs. Active (SPIVA) reports, consistently show that the majority of actively managed funds underperform their benchmark index over periods of ten years or more after fees, and that past outperformance does not reliably predict future outperformance. The manager who beat the market for five years is not much more likely to beat it for the next five years than a manager chosen at random. Fund management, at the aggregate level, looks much more like an independent process than a dependent one. Short-run outperformance is largely consistent with random variation.

This does not mean no skilled fund managers exist. It means that identifying them in advance, from a track record alone, is harder than it appears, because genuine skill and good luck produce indistinguishable short-run results. The difficulty is distinguishing signal from noise in a short time series.

In sport, coaches who pull players from a “cold” shooter or give the ball to a “hot” one are making decisions based on the hot hand belief. If the effect is real but small, the strategy may sometimes be right but frequently wastes the resources of context and attention. In any domain where “on a roll” narratives drive decisions, the question worth asking is: is this process actually autocorrelated, and do I have enough data to tell?

How to Spot It

The hot hand belief and the gambler’s fallacy can both be wrong, and they pull in opposite directions. The hot hand says: keep going, it will continue. The gambler’s fallacy says: stop, it must reverse. Neither is always right. Both are responses to streaks that ignore the fundamental question: what kind of process is this?

The tell for illegitimate hot hand reasoning is post-hoc selection. You have been watching a game, a portfolio, or a track record, and you noticed that things have gone well recently. The streak is visible in hindsight. The question is whether that visible streak tells you anything predictive.

For the hot hand belief to be valid, you need two things. First, an underlying mechanism: a reason to believe that the process is actually dependent, that early success genuinely raises the probability of subsequent success through some causal pathway. Second, enough data in the right form to detect that dependence statistically.

When those two conditions are absent, the “on a roll” narrative is almost always a post-hoc story imposed on noise. The streak that prompted the narrative was real. The causal structure you are inferring from it was not.

In fund management, the mechanism is weak. Capital allocation skill, even if real, operates in a market where everyone else is also trying to be clever, and where edge degrades rapidly as it becomes known. The short-run outperformance of a good manager is statistically hard to separate from the short-run outperformance of a lucky one. Five years of above-average returns, in a universe of thousands of funds, will reliably produce several funds with that track record even if no one has any skill at all. That is what random variation looks like.

The right response to a claimed hot hand is always to ask: what mechanism would produce dependence here, and how long a sequence would I need to confidently detect it?

Your Challenge

A fund manager has outperformed a broad market index in each of the past five consecutive years. The fund is now being actively marketed to retail investors on the basis of this track record.

Assuming that fund management over a five-year window is a roughly independent process, estimate the probability of five consecutive years of outperformance occurring by chance alone. Use the simplest possible assumption: each year has a 50 per cent chance of outperformance regardless of prior years.

Then consider: if there are five thousand actively managed funds available to UK retail investors, how many would you expect to have a five-year outperformance streak by chance alone, even if no fund has any skill? What does this tell you about the evidential value of the track record being presented to you?

Finally: what mechanism would you need to identify, and what additional data would you want, before the five-year record became meaningful evidence of skill rather than luck?

There is no answer on this page.

References

Original study: Gilovich, T., Vallone, R., and Tversky, A., “The hot hand in basketball: On the misperception of random sequences,” Cognitive Psychology, 17(3), 295—314 (1985). Full text available at: https://home.cs.colorado.edu/~mozer/Teaching/syllabi/7782/readings/gilovich%20vallone%20tversky.pdf

Reanalysis: Miller, J.B. and Sanjurjo, A., “Surprised by the Hot Hand Fallacy? A Truth in the Law of Small Numbers,” Econometrica, 86(6), 2019—2047 (2018). Published version: https://onlinelibrary.wiley.com/doi/abs/10.3982/ECTA14943. Working paper (open access): https://arxiv.org/abs/1902.01265

Accessible explanation of the streak selection bias: Data Colada post 88, “The Hot-Hand Artifact for Dummies and Behavioral Scientists,” https://datacolada.org/88

Overview of the debate with further citations: Wikipedia, “Hot hand,” https://en.wikipedia.org/wiki/Hot_hand

Fund performance persistence evidence: S&P Dow Jones Indices, SPIVA US Scorecard (published semi-annually). Summary findings available at: https://www.spglobal.com/spdji/en/research-insights/spiva/

Kahneman, D., Thinking, Fast and Slow, Farrar, Straus and Giroux (2011), Chapter 10. Covers the hot hand, the clustering illusion, and the original Gilovich et al. findings. Note that the book predates the Miller and Sanjurjo reanalysis and reflects the pre-2018 consensus.