Decision Theory
Von Neumann Minimax Theorem
In finite two-person zero-sum games, max-min equals min-max. The fundamental theorem of game theory, its proof via linear programming duality, and connections to adversarial robustness, GANs, and online learning.
Prerequisites
Why This Matters
The minimax theorem (von Neumann, 1928) is the foundational result of game theory. It states that in two-person zero-sum games with mixed strategies, the order of optimization does not matter: the maximizer's guarantee equals the minimizer's guarantee. This is not obvious. In general, (weak duality always holds), but equality (strong duality) requires structure.
The theorem appears throughout ML under different names. The GAN objective is a minimax problem. Adversarial robustness is formulated as . Online learning regret bounds rely on minimax arguments. LP duality, which underlies much of optimization theory, is a special case. Understanding minimax duality is prerequisite to understanding any of these.
Setup: Zero-Sum Games
Two-Person Zero-Sum Game
A two-person zero-sum game is defined by a payoff matrix . Player 1 (the row player) chooses a row . Player 2 (the column player) chooses a column . Player 1 receives and player 2 receives .
Player 1 wants to maximize . Player 2 wants to minimize it.
Mixed Strategies in Zero-Sum Games
Player 1 chooses a mixed strategy . Player 2 chooses . The expected payoff to player 1 is:
Player 1 maximizes this bilinear form. Player 2 minimizes it.
Maximin and Minimax Values
The maximin value is the best guarantee player 1 can secure:
The minimax value is the least that player 2 can hold player 1 to:
Weak duality always holds: . The minimax theorem states .
The Minimax Theorem
Von Neumann Minimax Theorem (1928)
Statement
For any matrix :
The common value is the value of the game. Optimal strategies achieving this value exist and form a Nash equilibrium of the zero-sum game.
Intuition
Consider the function . It is linear (hence convex) in for fixed , and linear (hence concave) in for fixed . The strategy spaces and are compact and convex. These three ingredients (convex-concave function, compact convex domains) are exactly what is needed for strong minimax duality. The key point: mixing strategies (randomizing) creates convexity, which closes the minimax gap.
Without mixed strategies, the maximin value in pure strategies can be strictly less than the minimax value. Matching Pennies has pure maximin value and pure minimax value , but the mixed value is .
Proof Sketch
Via LP duality. Player 1's maximin problem is:
This is a linear program. Write it in standard form:
The dual LP is player 2's minimax problem:
which is equivalent to .
By LP strong duality (which holds since both primal and dual are feasible, as the simplices are nonempty), the optimal values are equal: . Therefore .
Why It Matters
The minimax theorem tells you that zero-sum games are "solved" in a strong sense. The value is unique, and optimal strategies can be computed in polynomial time (by solving an LP). This is in sharp contrast to general Nash equilibria, which are PPAD-complete to compute.
In ML, the theorem justifies the GAN objective: the generator can achieve the data distribution at the Nash equilibrium of the minimax game. It also justifies adversarial training: the robust model exists because the minimax equality holds (under appropriate convexity conditions on the loss landscape).
Failure Mode
The theorem requires mixed strategies. In pure strategies, in general (e.g., Matching Pennies). The theorem also requires finite strategy spaces (or the appropriate generalization to compact convex sets with a convex-concave objective). For general functions on general domains, the minimax equality can fail.
Sion's Generalization
Sion Minimax Theorem (1958)
Statement
Let be a compact convex subset of a topological vector space and a convex subset. If satisfies:
- is quasiconvex and lower semicontinuous on for each
- is quasiconcave and upper semicontinuous on for each
Then:
Intuition
Sion's theorem extends von Neumann's result from bilinear functions on simplices to convex-concave (or more generally, quasiconvex-quasiconcave) functions on arbitrary compact convex sets. Only one of the two sets needs to be compact. This is the version most useful in continuous optimization and ML, where the strategy spaces are not simplices but arbitrary convex parameter spaces.
Proof Sketch
The proof uses the intersection property of compact convex sets (a topological fixed-point argument). For each , the sets (for varying ) have the finite intersection property by the quasiconcavity of in . Compactness of gives a point in the intersection, and taking yields the result.
Why It Matters
Sion's theorem is the version you actually use in ML proofs. The GAN minimax equality (for the population objective with infinite-capacity function classes), adversarial training with perturbation sets, and distributionally robust optimization all invoke Sion's theorem (or the closely related Fan minimax theorem) rather than von Neumann's original result.
Failure Mode
Both compactness and convexity-concavity are required. If the function is not convex-concave (e.g., on ), the minimax equality can fail. If neither set is compact (e.g., on ), the sup and inf may not be attained.
Saddle Points
Saddle Point
A pair is a saddle point of on if:
Equivalently, minimizes and maximizes . A saddle point exists if and only if the minimax equality holds and the optimizers are attained.
The minimax theorem guarantees that the zero-sum game has a saddle point in mixed strategies. At the saddle point, neither player can improve unilaterally: player 1 cannot increase the payoff, and player 2 cannot decrease it. This is exactly the Nash equilibrium condition for zero-sum games.
Computing Minimax Strategies
For a finite zero-sum game with matrix , the minimax strategies are computed by solving the LP:
This LP has variables and constraints (plus nonnegativity). It can be solved in time using interior-point methods. For games, the solution can be written in closed form using the indifference principle.
For continuous games (Sion's setting), minimax strategies are found by convex optimization methods: projected gradient descent-ascent, extragradient methods, or mirror descent. Convergence rates are for the duality gap.
Applications in Machine Learning
GANs
The GAN objective is:
For fixed generator and discriminator function classes of sufficient capacity, this is a minimax problem. The minimax theorem (Sion's version) applies when the inner maximization over is concave (it is, in ) and the outer minimization over is convex in the relevant sense. At the saddle point, .
Adversarial Robustness
Adversarial training solves:
The inner maximization finds the worst-case perturbation. The outer minimization trains the model to be robust. If is convex in and the perturbation set is compact and convex, the minimax theorem guarantees the order of optimization does not matter.
Online Learning and Regret
The minimax regret framework in online learning asks: what is the best worst-case regret an algorithm can achieve against an adversary? For the experts problem, the minimax optimal regret is for rounds and experts. This is proved using a minimax argument: the algorithm and adversary play a zero-sum game where the payoff is regret.
Common Confusions
Minimax equality requires convexity-concavity, not just any function
For a general function , the weak inequality always holds. Equality requires the function to be convex-concave (or quasiconvex-quasiconcave with appropriate compactness). The neural network parameterization in GANs makes the actual optimization landscape non-convex-concave, which is why GAN training can fail to converge despite the population-level minimax theorem applying.
The minimax theorem says nothing about algorithms
The theorem guarantees that the value exists and optimal strategies exist. It does not say that gradient descent-ascent converges to the saddle point. Simultaneous gradient descent on a bilinear function can cycle forever without converging. Extra-gradient methods, optimistic gradient descent, or two-time-scale updates are needed for convergence.
LP duality and game-theoretic minimax are the same theorem
The proof of the minimax theorem via LP duality and the proof of LP strong duality via Farkas' lemma are logically equivalent. Any one implies the other. This is not a coincidence: solving a zero-sum game IS solving a linear program, and vice versa. Von Neumann proved the minimax theorem in 1928; Dantzig formulated the LP in 1947 and recognized the connection.
Exercises
Problem
For the zero-sum game with payoff matrix , find the value of the game and the optimal mixed strategies for both players.
Problem
Prove the weak duality inequality: for any and :
Conclude that .
Problem
Write the LP for computing the maximin strategy of a zero-sum game with payoff matrix . Write its dual LP. Verify that the dual corresponds to the minimax strategy for player 2. What is the relationship between the number of variables and constraints in the primal and dual?
Problem
Show that Sion's minimax theorem implies Von Neumann's minimax theorem as a special case. Verify that the bilinear function on satisfies all hypotheses of Sion's theorem.
References
Canonical:
- Von Neumann, "Zur Theorie der Gesellschaftsspiele" (1928), Mathematische Annalen
- Von Neumann & Morgenstern, Theory of Games and Economic Behavior (1944), Chapter 3
- Sion, "On General Minimax Theorems" (1958), Pacific Journal of Mathematics
Optimization:
- Boyd & Vandenberghe, Convex Optimization (2004), Chapter 5 (LP duality)
- Bertsimas & Tsitsiklis, Introduction to Linear Optimization (1997), Chapter 4
ML Applications:
- Goodfellow et al., "Generative Adversarial Nets" (2014), Section 4
- Madry et al., "Towards Deep Learning Models Resistant to Adversarial Attacks" (2018)
- Cesa-Bianchi & Lugosi, Prediction, Learning, and Games (2006), Chapters 4, 7
Next Topics
Building on the minimax theorem:
- Adversarial machine learning: minimax formulations for robustness
Last reviewed: April 2026
Prerequisites
Foundations this topic depends on.
- Game Theory FoundationsLayer 2
- Common Probability DistributionsLayer 0A
- Sets, Functions, and RelationsLayer 0A
- Basic Logic and Proof TechniquesLayer 0A
- Convex Optimization BasicsLayer 1
- Differentiation in RnLayer 0A
- Matrix Operations and PropertiesLayer 0A
- Convex DualityLayer 0B