The Curriculum

44 units. One per day, fifteen minutes each. Ten weeks to statistical self-defence. Follow the sequence or jump to any unit from the reference library.

★ Essential Tier 1 — highest Pareto return
◆ Powerful Tier 2 — strong return
⬡ Bayes Bayes thread
Level 0 · 5 units
Zero-Knowledge On-Ramp
No prior knowledge required. Five units to build the basic vocabulary.
0.1
Numbers at Scale
10 min

A million and a billion sound like cousins. They are not. Most numerical manipulation in public life depends on you not feeling the difference. This unit corrects that, permanently.

0.2
Percentages, Fractions, and Ratios
0 15 min

The vocabulary of proportion. Percentages, fractions, and ratios are the building blocks of almost every statistical claim — and the source of some of the most reliably effective manipulation. This unit covers the percentage point versus percentage change distinction, what 'X times more likely' actually means, and why every percentage needs a denominator.

0.3
The Three Averages: First Contact
★ Essential 12 min

The word 'average' conceals three different calculations, each giving a different answer. This unit shows why mean and median diverge in skewed distributions — and how choosing between them is one of the most common tools of numerical manipulation.

0.4
Reading Graphs Critically
★ Essential 12 min

A graph communicates through visual impression before you process the numbers. That impression can be engineered. Five tricks — the truncated y-axis, the compressed x-axis, area distortion, dual axes, and the missing baseline — account for the vast majority of graph manipulation you will encounter.

0.5
What Is Risk? The Intuitive Primer
★ Essential 12 min

Before formal probability, you need an intuitive vocabulary for risk. This unit covers absolute risk, baselines, and why your gut feeling about risk is systematically, predictably wrong — and how that gets exploited.

Area 1 · 7 units
Foundations
Probability, uncertainty, and the laws of chance.
1.1
What Is Probability?
12 min

Two legitimate interpretations of probability, both useful, often confused. The frequentist view treats probability as long-run frequency; the Bayesian view treats it as a degree of belief. The distinction matters every time you read a forecast, a poll, or a risk figure in public life.

1.2
What Is Randomness?
12 min

Most people have a wrong model of what randomness looks like. Random sequences contain runs and clusters as a matter of course. Understanding this protects you from one of the most seductive errors in statistics: seeing a pattern in noise and inventing a cause for it.

1.3
The Laws of Probability
◆ Powerful 12 min

The four formal rules that govern how probabilities combine — addition for exclusive events, addition for non-exclusive events, multiplication for independent events, and the complementary rule. Plus the specific way these rules get weaponised to mislead.

1.4
Conditional Probability
★ Essential 15 min

The probability of an event changes depending on what you already know. P(A|B) is not the same as P(B|A), and confusing the two has sent innocent people to prison and led doctors to badly misread test results. This unit gives you the tool that sits at the heart of Bayes' theorem.

1.5
Expected Value: The Number Behind Every Bet, Ticket, and Policy
◆ Powerful 12 min

What probability theory predicts will happen on average, over many trials. How to calculate it, when to use it, when it misleads you, and why the people selling you lottery tickets and insurance policies already know it and are counting on you not to.

1.6
Variance as Spread
◆ Powerful 12 min

Two data sets with identical averages can be entirely different in character. Variance and standard deviation measure how far outcomes scatter around the centre — and in finance, medicine, and public policy, ignoring spread is how people get badly misled.

1.7
First Encounter with Bayes' Theorem
◆ Powerful 15 min

The most important single idea in the curriculum, introduced intuitively. A positive test result for a rare disease is probably wrong, even when the test is 99% accurate. Understanding why requires prior probability, likelihood, and posterior belief — the three ingredients of Bayesian reasoning.

Area 2 · 11 units
Measurement and Its Limits
The formal toolkit — sampling, distributions, confidence, p-values.
2.1
Descriptive Statistics: The Full Picture
◆ Powerful 12 min

Mean and median were the start. This unit adds variance, standard deviation, percentiles, quartiles, the five-number summary, box plots, and weighted averages — the full toolkit for understanding what a dataset actually looks like, and for noticing when someone has shown you only part of it.

2.2
The Normal Distribution
15 min

Why the bell curve appears everywhere, what it actually implies, and what happens when the people running your pension fund assume it holds when it does not. The 68-95-99.7 rule, z-scores, and the normality assumption that helped cause the 2008 financial crisis.

2.3
When Distributions Are Not Normal
15 min

Real data is rarely bell-shaped. Skew, fat tails, bimodal distributions, log-normal distributions, and power laws each produce systematic errors when treated as if they were normal. This unit shows what those errors look like and why they cost people money, elections, and lives.

2.4
Sampling: What a Study Tells You
★ Essential 12 min

Every study is a sample. The quality of the conclusion depends on the quality of the sample. Population vs sample, random sampling, convenience sampling, self-selection bias, and stratified sampling — with the 1936 Literary Digest disaster as the object lesson.

2.5
Sample Size: How Many Is Enough?
12 min

A thousand randomly chosen people can accurately represent sixty million. A million carefully selected people might tell you almost nothing. Sample size matters, but size alone is not the point. How the sample was gathered is what determines whether the number is meaningful.

2.6
Confidence Intervals — What the Uncertainty Band Actually Says
◆ Powerful 15 min

Almost everyone who reads a confidence interval misreads it. The misreading is not a minor technicality. It is the difference between knowing something is uncertain and believing it is not.

2.7
P-Values: The Most Abused Number in Science
◆ Powerful 15 min

The p-value is the most cited and most misunderstood number in public science. Almost everyone who uses it is using it wrong, including many scientists. Here is what it actually means, why the 0.05 threshold is completely arbitrary, and why a result can be both statistically significant and completely pointless.

2.8
Type I and Type II Errors
12 min

Every statistical test can fail in exactly two ways. Which failure you find more tolerable is not a technical question. It is a values question, and the answer determines how the system treats you.

2.9
Hypothesis Testing: The Full Framework
15 min

The complete null-hypothesis significance testing procedure, assembled from confidence intervals, p-values, and error types. One-tailed versus two-tailed tests, degrees of freedom, multiple comparisons, and the one trick that inflates false discoveries without anyone technically cheating.

2.10
Bayes vs Null-Hypothesis Testing
◆ Powerful 15 min

The two dominant frameworks in statistics answer completely different questions. Null-hypothesis testing tells you how surprising your data would be if nothing were going on. Bayesian inference tells you how probable your hypothesis is, given what you now know. These are not the same thing, and the difference matters every time a drug is approved, a trial proceeds to verdict, or a screening programme is designed.

2.11
The Replication Crisis
◆ Powerful 15 min

Most published scientific findings do not replicate. This is not a failure of individual scientists — it is a structural consequence of how science is conducted and reported. Here is what it means, why it happened, and how to read a single study properly.

Area 3 · 12 units
Applied Detection
Twelve tricks used in the wild, with real examples and detection tests.
3.1
The Relative Risk Trick
★ Essential 15 min

The single most important unit in the curriculum. The most common trick in medicine, policy, and advertising. A treatment that reduces risk from 2% to 1% is '50% effective' — but reduces your absolute risk by one percentage point.

3.2
Cherry-Picking — Time Periods, Subgroups, and Studies
★ Essential 15 min

Any dataset contains patterns. The selection of which patterns to present is a powerful tool. This unit covers cherry-picked time periods, subgroup analyses, and selective citation of studies.

3.3
The Misleading Graph — A Comprehensive Guide
★ Essential 15 min

Every trick in the visual statistics arsenal: the truncated axis, the inverted axis, the dual axis, the area distortion, and the missing uncertainty band. With documented real-world examples from governments, media, and financial advertising.

3.4
Survivorship Bias
★ Essential 12 min

We only see the planes that came back. Abraham Wald noticed this in 1943 and saved bomber crews' lives. The same blind spot distorts investment returns, entrepreneurship mythology, and medical evidence — wherever failures disappear from the data before you see it.

3.5
Regression to the Mean
◆ Powerful 12 min

Extreme measurements are followed by less extreme ones — not because anything changed, but because that is how randomness works. Misattributing this statistical law to an intervention is one of the most common errors in medicine, management, and sport.

3.6
The Unrepresentative Sample
★ Essential 15 min

The politics of who gets studied and who does not. WEIRD psychology populations, male-dominant clinical trials, the Tuskegee legacy, volunteer bias, and attrition bias — and how conclusions drawn from one group get quietly applied to everyone else.

3.7
Correlation Is Not Causation
★ Essential 15 min

The most frequently stated statistical fact and the least frequently applied one. Three structural reasons why correlation does not imply causation — confounding, reverse causation, and chance — with the Bradford Hill criteria as a framework for when it might.

3.8
The Ecological Fallacy
◆ Powerful 12 min

Using statistics about groups to draw conclusions about individuals is a structural error with serious real-world consequences — in criminal profiling, public health policy, and cross-national research. Here is how to recognise it and its mirror image.

3.9
The Prosecutor's Fallacy
◆ Powerful 15 min

Confusing P(evidence|innocent) with P(innocent|evidence). This error has contributed to wrongful convictions. The Sally Clark case, DNA database searches, and how to spot the tell in any courtroom probabilistic argument.

3.10
P-Hacking and Data Dredging
◆ Powerful 15 min

Statistical significance can be manufactured without fraud, simply by making enough analysis choices until the right number appears. This is p-hacking, and it is built into how academic publishing works. Understanding it changes how you read every headline that says 'scientists find'.

3.11
Publication Bias and the File Drawer Problem
◆ Powerful 15 min

The studies that get published are not a random sample of all studies conducted. Positive findings are published; negative findings are buried. The result is a scientific literature systematically tilted toward conclusions that support the people funding the research — and a public making decisions based on a distorted picture.

3.12
Risk Communication: The Full Toolkit
◆ Powerful 15 min

The synthesis unit. NNT, NNH, absolute vs relative risk, natural frequencies, icon arrays, sensitivity, specificity, PPV, NPV, and how PPV collapses when you screen a low-prevalence population. The HIV test worked through in full. The mammography debate as a case study in honest versus dishonest risk communication.

Area 4 · 9 units
The Anti-Curriculum
Cognitive failure modes and their corrections. Ends with Bayes as the synthesis.
4.1
The Availability Heuristic
◆ Powerful 12 min

We judge the probability of an event by how easily examples come to mind. Vivid, recent, or emotionally charged events are overweighted. This one cognitive shortcut shapes public policy, insurance markets, and personal decisions in ways that have nothing to do with actual risk.

4.2
Base Rate Neglect
★ Essential 15 min

The most consequential cognitive error in everyday probabilistic reasoning. When a specific piece of evidence arrives, the mind discards the background frequency of the event in the population. The cab problem shows you exactly how this happens, and Bayes' theorem shows you exactly how to stop it.

4.3
The Conjunction Fallacy
◆ Powerful 12 min

People consistently judge a specific, detailed scenario as more probable than a broader one that contains it. This violates a foundational rule of probability. Understanding why it happens, and what it looks like in the wild, is the protection.

4.4
The Gambler's Fallacy
◆ Powerful 12 min

Past random events are taken to influence future random events, even when there is no causal mechanism connecting them. The casino industry is built, in part, on this error. Understanding it — really understanding it, not just nodding at it — changes how you read any streak.

4.5
The Hot Hand Fallacy: And Why It Is Different
◆ Powerful 13 min

The belief that a player on a hot streak will keep performing well was declared a cognitive illusion in 1985. Then in 2018 a mathematician found a flaw in the original analysis. The hot hand may be real after all. This unit covers what changed, why it matters, and how it connects to a deeper point about dependent processes.

4.6
Anchoring and Numerical Estimation
◆ Powerful 12 min

The first number you encounter shapes every estimate that follows, even when you know that number was chosen at random. Anchoring operates in salary negotiations, retail pricing, and courtroom sentencing. Understanding it changes how you read any situation where someone else names a number first.

4.7
Overconfidence and Calibration
◆ Powerful 13 min

People's confidence in their own judgements reliably exceeds their accuracy. This is not a personality flaw — it is a systematic feature of how human minds form predictions. Understanding it, and the correctives that actually work, is the difference between someone whose plans routinely collapse and someone who consistently delivers.

4.8
The Representativeness Heuristic
◆ Powerful 13 min

Judgements of probability are replaced by judgements of similarity to a prototype. The conjunction fallacy, base rate neglect, and the small sample fallacy are all the same mistake wearing different clothes. When you understand the single mechanism behind all three, you become much harder to fool.

4.9
Bayes as the Synthesis: The Calibrated Mind
◆ Powerful 15 min

The final unit. Bayes' theorem is not just a calculation — it is a description of how a rational mind should work. Every cognitive failure in this area is, at bottom, a failure to apply Bayes. Calibration. Superforecasting. Prediction markets. And the three questions that, if you ask them habitually, will change how you think.