Skip to main content

Comparison

Central Limit Theorem vs. Law of Large Numbers

The LLN says the sample mean converges to the population mean. The CLT says how fast and in what shape. Different assumptions (finite mean vs. finite variance), different scaling (n vs. sqrt n), different conclusions (consistency vs. fluctuation shape).

Last reviewed: May 12, 2026

What Each Answers

Both results are about the sample mean Xˉn=1ni=1nXi\bar{X}_n = \frac{1}{n}\sum_{i=1}^n X_i of i.i.d. random variables, but they answer different questions.

Law of Large Numbers. Does Xˉn\bar{X}_n converge to E[X]\mathbb{E}[X]? The LLN says yes, under finite-mean i.i.d. assumptions. It is a zero-th-order result: it identifies the limit, nothing more.

Central Limit Theorem. What is the limiting shape of the fluctuations XˉnE[X]\bar{X}_n - \mathbb{E}[X] as nn \to \infty? Under additional finite-variance assumptions, the CLT says they are Gaussian after rescaling by n\sqrt{n}. It is a first-order result: it identifies the limit and the order of fluctuations around it.

Together they form the two-step asymptotic picture: LLN says where the estimator is going, CLT says how fast it gets there and what the remaining randomness looks like.

Side by Side

AspectLLNCLT
StatementXˉnμ\bar{X}_n \to \mun(Xˉnμ)dN(0,σ2)\sqrt{n}(\bar{X}_n - \mu) \xrightarrow{d} \mathcal{N}(0, \sigma^2)
Convergence modeIn probability (WLLN) or almost sure (SLLN)In distribution
Minimal assumption (i.i.d.)EX<\mathbb{E}\lvert X \rvert < \inftyE[X2]<\mathbb{E}[X^2] < \infty
ScalingNone (the sample mean itself)n\sqrt{n}
Information returnedThe limitThe limit + the fluctuation shape + the rate
Confidence-interval useJustifies consistencyConstructs the interval
Fails forCauchy, Pareto α1\alpha \leq 1Pareto α2\alpha \leq 2, all α\alpha-stable laws with α<2\alpha < 2
Replacement when assumption failsNone: estimator inconsistentGeneralized CLT: α\alpha-stable limit

The LLN-vs-CLT distinction is not "weak result vs. strong result". They require different assumptions and produce different kinds of information. A distribution can satisfy the LLN but fail the CLT (Pareto, α(1,2)\alpha \in (1, 2)); in that case the sample mean still converges to the right limit, but not at the Gaussian 1/n1/\sqrt{n} rate.

The Practical Question: Which Do I Use?

Most undergraduate statistics blurs the two. The blur is harmless when the underlying distribution has both finite mean and finite variance. For more general work the distinction matters.

Use the LLN when:

  • The question is "is my estimator consistent?" or "does the sample mean converge to something?"
  • You are constructing a Monte Carlo estimator and want a correctness guarantee.
  • You are proving consistency of an estimator like maximum likelihood or empirical risk minimization, where the first-order statement is enough.

Use the CLT when:

  • The question is "how confident am I that my estimate is close to the truth?", and the answer needs a confidence interval.
  • You need a finite-nn rate of 1/n1/\sqrt{n}, e.g. for empirical risk fluctuations around population risk.
  • You are calibrating a hypothesis test that uses a Gaussian critical value.
  • You are reasoning about the asymptotic normality of an MLE or M-estimator.

A Worked Contrast

Bernoulli sampling. Toss a coin with bias pp for nn trials, observe Xˉn\bar{X}_n.

LLN says: Xˉnp\bar{X}_n \to p almost surely. The fraction of heads settles at the bias.

CLT says: n(Xˉnp)dN(0,p(1p))\sqrt{n}(\bar{X}_n - p) \xrightarrow{d} \mathcal{N}(0, p(1-p)). The fluctuations are Gaussian with explicit variance.

LLN alone tells you nothing about how many tosses you need to verify p=0.5p = 0.5 vs p=0.51p = 0.51. CLT plus the Berry-Esseen bound tells you: the standard error of Xˉn\bar{X}_n is p(1p)/n\sqrt{p(1-p)/n}, and the Gaussian approximation is good to within O(1/n)O(1/\sqrt{n}) in sup-distance, so for n=104n = 10^4 you need to see Xˉn0.51.6/104=0.016\bar{X}_n - 0.5 \approx 1.6/\sqrt{10^4} = 0.016 to declare a significant difference from a fair coin at the 5% level.

A confidence interval needs the CLT. A consistency proof needs the LLN. They are different jobs.

The Heavy-Tail Wedge

The two thresholds (finite mean for LLN, finite variance for CLT) define three regimes for i.i.d. samples. The middle one is where the two theorems split.

Distribution tailLLNCLTPractical answer
Finite variance (Gaussian, bounded, Pareto α>2\alpha > 2)HoldsHoldsBoth apply: sample mean converges Gaussian-fast.
Finite mean but infinite variance (Pareto 1<α21 < \alpha \leq 2)HoldsFailsSample mean still converges to the right value, but at rate n11/αn^{1-1/\alpha} not n\sqrt{n}, and the limit law is α\alpha-stable, not Gaussian. See LLN-and-CLT-failures-under-heavy-tails.
Infinite mean (Cauchy, Pareto α1\alpha \leq 1)FailsFailsSample mean is inconsistent. Use the median or a trimmed mean.

The middle regime is the one most often overlooked in introductory courses, and it is where many real applications (financial returns, insurance losses, word frequencies) sit.

Common Confusions

"The CLT proves the LLN." No. The CLT requires finite variance; the LLN does not. The LLN under only finite mean is a strictly weaker statement that the CLT cannot derive directly. If you have a sequence with infinite variance but finite mean, the LLN applies and the CLT does not.

"The LLN is just the n = ∞ limit of the CLT." The LLN says the sample mean has a degenerate limit at μ\mu. The CLT says the fluctuations around that limit (after n\sqrt{n} rescaling) have a non-degenerate Gaussian shape. These are different objects, not two views of the same thing.

"For large enough nn, the sample mean is approximately Gaussian." Only if the CLT applies. If Var(X)=\mathrm{Var}(X) = \infty, no nn is large enough; the sample mean has a non-Gaussian limit law, and confidence intervals built on the Gaussian approximation are wrong by an asymptotically growing factor.

Quick Decision Rule

Want a limit? Use LLN. Want a rate or a confidence interval? Use CLT. Heavy tails? Check which regime you are in, then use the appropriate theorem (or its replacement for the failure case).

References

Canonical:

  • Durrett, Probability: Theory and Examples (5th ed., 2019), Chapter 2 (LLN) and Chapter 3 (CLT).
  • Billingsley, Probability and Measure (3rd ed., 1995), Sections 6 and 27.

Current:

  • van der Vaart, Asymptotic Statistics (1998), Chapter 2 (LLN and CLT in the context of statistical estimation).
  • Wainwright, High-Dimensional Statistics: A Non-Asymptotic Viewpoint (2019), Chapter 2 (finite-sample concentration as the non-asymptotic counterpart to CLT).