Foundations
Lognormal Distribution
A random variable is Lognormal if its logarithm is Normal. The density, mean, variance, median, and mode all have closed forms in the two underlying Normal parameters. The Lognormal is the multiplicative analogue of the Normal: a product of many independent positive factors is approximately Lognormal in the same way a sum is approximately Normal. Applications cover financial returns (with the heavy-tail caveat that real returns are heavier than Lognormal), particle sizes, lifetimes, and insurance severity.
Prerequisites
Plain-Language Definition
A random variable is Lognormal if its logarithm is Normal. Take any Normal random variable with mean and variance , and define . The distribution of is Lognormal. Equivalently, is Lognormal whenever is Normal.
The Normal arises as the limit of sums of independent random variables (the central limit theorem). The Lognormal arises as the limit of products of independent positive random variables, because the logarithm of a product is a sum. Anywhere multiplicative compounding is more natural than additive (asset returns, particle size after repeated random splits, biological growth), the Lognormal is a more honest baseline than the Normal.
Definition
Lognormal Distribution
A positive random variable has a Lognormal distribution with parameters and when . Equivalently, the density of is
The parameters and are the mean and variance of , not of itself.
The density vanishes at zero, rises to a single mode, and decays super-polynomially in the right tail. The shape parameter is ; the location of the mode and the spread of the upper tail both stretch as grows. As the distribution concentrates near and becomes approximately Normal in a local sense.
Why This Matters
The Lognormal is the natural multiplicative analogue of the Normal. Three places where this matters in practice.
-
Multiplicative processes. Asset returns compounded over periods are products of independent (or near-independent) gross returns. Under modest assumptions about the per-period log-return distribution, the central limit theorem applied to the log of the product gives an approximately Lognormal price after many periods. This is the formal basis of Black-Scholes option pricing and any geometric-Brownian-motion model.
-
Right-skewed positive data. Particle sizes, drug-trial responses, household incomes, file sizes, insurance claim severities. All are nonnegative and right-skewed, and many fit a Lognormal model better than a Normal at the cost of a single nonlinear transformation.
-
Survival and reliability work. The Lognormal hazard rises and then falls, which is a useful shape for failure-time data where early-life and end-of-life failures dominate but the middle of the life has comparatively few failures. Weibull is a more common reliability default, but the Lognormal is the standard alternative when the Weibull hazard shape is wrong.
The classical warning is that financial returns are heavier-tailed than Lognormal at long horizons. The empirical excess kurtosis of equity returns is decades of literature; a Lognormal model gets the bulk right and underestimates tail risk. For insurance severity, Pareto and Weibull are the standard alternatives once the data shows a heavier tail than Lognormal can support.
Moments
Lognormal Mean, Variance, Median, Mode
Statement
Intuition
The median is the simplest of the four: is Normal with median , and the monotone exponential preserves quantiles, so has median . The mean is strictly larger because of the convexity of the exponential and Jensen's inequality: , and the correction is exactly . The mode is strictly smaller because the density is right-skewed.
Proof Sketch
Use the moment generating function of a Normal: for . Setting gives . Setting gives , so . For the mode, differentiate the density, set to zero, and solve .
Why It Matters
Forgetting the correction in the mean is one of the most common errors in applied work. If you fit a Normal to log-returns and then exponentiate the estimated log-mean, you get the median of returns, not the mean. The two differ by the multiplicative factor , which for a typical equity sigma of is about over a year and grows quadratically with horizon.
Failure Mode
The MGF of a Lognormal is infinite for every . The Lognormal does not have an MGF in the usual sense, and identities that depend on MGF uniqueness do not apply. The Lognormal still has moments of every order, but the moment sequence does not uniquely determine the distribution; there is a Stieltjes moment-determinacy failure that produces non-identifiable density modifications.
The Multiplicative Central Limit Theorem
Product of Independent Positive r.v.s converges to Lognormal
Statement
Let be iid positive random variables with and . Then for the product ,
Equivalently, is approximately Lognormal with parameters and for large .
Intuition
This is the central limit theorem applied to . The product of many independent factors is Lognormal in the same way a sum of many independent terms is Normal. The Normal limit for the log is exact in the limit; the Lognormal claim for the product is its exponentiated counterpart and is approximate in the same sense.
Proof Sketch
Define . The are iid with finite mean and finite variance by assumption, so the classical central limit theorem applied to the gives the displayed convergence in distribution. Continuous mapping under the exponential transforms convergence of the log-sum into convergence in distribution of the product to a Lognormal.
Why It Matters
The result formalizes the intuition that multiplicative compounding produces Lognormal aggregates. Geometric Brownian motion in finance is exactly this limit when the time step shrinks and the number of independent multiplicative shocks grows. The same logic gives Lognormal-shaped distributions for biological growth and for repeated-fragmentation particle sizes.
Failure Mode
The result requires finite variance of . If have a heavy enough left tail (e.g. can be very close to zero), has infinite variance and the Lognormal limit fails. Power-law tails in also break finite-variance assumptions and can produce stable-law limits for the product instead of Lognormal.
Worked Example: Median vs Mean of Equity Returns
A daily equity log-return has approximately and (about 19 percent annualized). Over trading days, the cumulative log-return is approximately , with .
The cumulative gross return is approximately Lognormal with parameters and . Therefore:
- , a typical annual gross return near percent.
- , an annual expected gross return near percent.
- , .
The gap between percent (median) and percent (mean) is the volatility drag. Quoting "average return" without specifying mean versus median misrepresents one or the other by close to basis points per year at this volatility level.
Common Misconceptions
The parameters mu and sigma are not the mean and variance of X
is the mean of , not of . The mean of is . Mistaking for is the most common applied error with this distribution, and it propagates through every downstream calculation.
The Lognormal MGF is not finite at any positive value
for every , because the density tail of decays sub-exponentially. Any inference procedure that relies on the existence of the MGF (Chernoff bounds, tilted measures, MGF uniqueness in the Stieltjes sense) breaks. The Lognormal still has all moments and a well-defined cumulant generating function on the imaginary axis, but the right-half-plane MGF is unavailable.
The Lognormal is not heavy-tailed in the regularly varying sense
The right tail of the Lognormal decays faster than every polynomial. It is heavier than Normal but lighter than any power law. For practical purposes, the Lognormal tail behaves like a slowly-decaying exponential in the log scale, and tail probabilities can be computed accurately from the Normal CDF on . Confusing Lognormal with Pareto leads to dramatic underestimates of extreme-quantile risk for genuinely power-law data.
Comparison: Normal vs Lognormal
The two are intimately related, and the natural way to choose between them is to think about whether shocks are additive or multiplicative on the variable of interest.
- Normal. Supported on the whole real line. Mean and variance are the parameters. Symmetric. Arises from additive aggregation. Right tail and left tail are equal.
- Lognormal. Supported on positive reals. Log-mean and log-variance are the parameters. Right-skewed. Arises from multiplicative aggregation. The mean is strictly greater than the median.
A useful test: simulate a thousand independent draws, take log, and check whether the resulting sample is symmetric. If yes, Lognormal is reasonable. If the log-sample is still skewed, neither Normal nor Lognormal fits well, and a heavier-tailed alternative (Pareto, Weibull, or a finite-mixture) is needed.
For a side-by-side severity-modeling treatment, see ActuaryPath's Lognormal page at https://www.actuarypath.com/concepts/lognormal-distribution/ , which approaches the same distribution from the loss-modeling angle (ASTAM and FAM-S, fitted by MLE, used as a parametric severity component in compound distributions).
Exercises
Problem
Let . Compute , , the median, the mode, and .
Problem
A particle's diameter is the product of independent random shrinkage factors, each iid Lognormal with and (a shrinkage of about 10 percent per step in log scale). Find the approximate distribution of where is the initial size.
Problem
An insurance loss is modeled as in dollars. Compute the expected loss, the standard deviation of the loss, the median, and the 95th percentile.
Problem
Show that the MLE of from an iid sample from a Lognormal distribution reduces to the MLE of the Normal mean and variance applied to .
Problem
Compute for in closed form, and interpret the result for (one standard log-deviation above the median).
Beyond Lognormal: When the Tail Is Heavier
The Lognormal tail decays faster than every polynomial but slower than any exponential. Real-world data sometimes has a tail heavier than Lognormal can support. Three standard moves when that happens:
- Fit a Pareto tail above a threshold and a Lognormal body below. This is the peaks-over-threshold approach from extreme-value theory.
- Fit a Weibull with shape parameter below 1, which gives a tail heavier than Lognormal but lighter than Pareto.
- Fit a finite mixture of Lognormals or a Lognormal-Pareto composite, accepting the loss of analytical tractability for a better tail fit.
Diagnostics: a log-log survival-function plot. Pareto-like data shows a straight line on a log-log plot. Lognormal data shows a curve that bends down. Weibull data with shape less than 1 sits between the two.
References
- Casella, G., and Berger, R. L. (2002). Statistical Inference, 2nd ed., Duxbury. Section 3.3 covers the Lognormal as a transformation of the Normal and lists the moment formulas.
- Blitzstein, J. K., and Hwang, J. (2019). Introduction to Probability, 2nd ed., Chapman and Hall / CRC. Chapter 5 includes the Lognormal in the catalog of continuous distributions with worked examples on financial returns.
- For the loss-modeling and severity-fitting perspective, see ActuaryPath's Lognormal page at https://www.actuarypath.com/concepts/lognormal-distribution/ and Klugman, Panjer, Willmot (2019), Loss Models: From Data to Decisions, 5th ed., Wiley, Chapter 5.
Last reviewed: May 12, 2026
Canonical graph
Required before and derived from this topic
These links come from prerequisite edges in the curriculum graph. Editorial suggestions are shown here only when the target page also cites this page as a prerequisite.
Required prerequisites
4- Common Probability Distributionslayer 0A · tier 1
- Distributions Atlaslayer 0A · tier 1
- Normal Distributionlayer 0A · tier 1
- Central Limit Theoremlayer 0B · tier 1
Derived topics
0No published topic currently declares this as a prerequisite.