RL Theory
Minimax and Saddle Points
Minimax theorems characterize when max-min equals min-max. Saddle points arise in zero-sum games, duality theory, GANs, and robust optimization.
Prerequisites
Why This Matters
Minimax problems appear whenever two objectives conflict. In zero-sum games, one player maximizes what the other minimizes. In robust optimization, you minimize loss under the worst-case perturbation. In GANs, the generator minimizes what the discriminator maximizes. The minimax theorem tells you when you can swap the order of max and min without changing the value. When this swap is valid, the problem has a saddle point and is tractable. When it fails, the problem is harder.
Mental Model
Consider a function where one player chooses to maximize and the other chooses to minimize. The max-player moves first in : they choose knowing the min-player will respond optimally. The min-player moves first in : they choose knowing the max-player will respond optimally. Moving second is always at least as good, so . The minimax theorem identifies conditions under which equality holds.
Core Definitions
Saddle Point
A point is a saddle point of if:
Equivalently: maximizes and minimizes . Neither player can improve by unilateral deviation.
Minimax Value and Maximin Value
The maximin value is .
The minimax value is .
The weak duality inequality always holds: . When , we say the minimax theorem holds and the common value is the value of the game.
Convex-Concave Function
A function is convex-concave if is convex for each fixed and is concave for each fixed . Equivalently, the max-player faces a concave objective (wants to maximize) and the min-player faces a convex objective (wants to minimize).
Main Theorems
Von Neumann Minimax Theorem
Statement
Under the above conditions:
Moreover, a saddle point exists.
Intuition
Convexity-concavity plus compactness ensures that neither player benefits from going first. The convex-concave structure prevents "misaligned" curvature that would let one player exploit knowledge of the other's choice. Compactness ensures the optima are attained (not just approached in the limit).
Proof Sketch
One proof uses the separating hyperplane theorem. Define . The convex-concave structure makes convex. If maximin minimax, there is a gap, and a separating hyperplane argument yields a contradiction. An alternative proof uses Brouwer's fixed point theorem on the best-response correspondence.
Why It Matters
This theorem is the foundation of game theory and connects directly to linear programming duality. In a finite zero-sum game (matrix game), the strategy spaces are simplices (compact, convex) and the payoff is bilinear (convex-concave). Von Neumann's theorem guarantees a value and optimal mixed strategies. Strong duality in LP is a corollary: the primal and dual LP values are equal.
Failure Mode
Without convexity-concavity, the minimax gap can be strict. For example, let (not convex, just two points) and . Then (for any , the min-player matches) but (for any , the max-player matches). The gap is 1. Convexifying the strategy sets (allowing mixed strategies) restores equality.
Without compactness, the optima may not be attained even when the function is convex-concave. Consider on : neither nor is well-defined.
Sion Minimax Theorem
Statement
Under these conditions:
Intuition
Sion's theorem relaxes Von Neumann's conditions in two ways: it requires quasiconvexity-quasiconcavity instead of convexity-concavity, and need not be compact (only must be compact). This covers cases where the min-player's domain is unbounded.
Proof Sketch
The proof uses a topological intersection argument. For each finite subset , the set is nonempty and closed by upper semicontinuity and quasiconcavity. The finite intersection property (from compactness of ) yields a point where for the critical value .
Why It Matters
Sion's theorem applies to infinite-dimensional settings (e.g., function spaces in variational problems) and to cases where strict convexity-concavity fails but quasiconvexity-quasiconcavity holds. It is the most general classical minimax result.
Failure Mode
Without compactness of at least one of the two sets, the theorem can fail even with quasiconvexity-quasiconcavity. Both sets being merely convex and unbounded is not enough.
Connection to Duality
Strong duality in linear programming is a special case of the minimax theorem. The LP primal-dual pair:
can be written as where is the Lagrangian. This is linear (hence convex-concave) in , and the constraint sets are convex. Strong duality () follows from the minimax theorem.
More generally, for convex programs: strong duality holds when Slater's condition is satisfied, and the saddle point of the Lagrangian corresponds to the primal-dual optimal pair.
Connection to GANs
The GAN objective is:
For fixed , the optimal discriminator is , and the resulting objective for is , where JSD is the Jensen-Shannon divergence. The minimax theorem does not directly apply because the generator and discriminator are parameterized by neural networks (non-convex). Training GANs is therefore a non-convex saddle-point problem, and convergence guarantees are limited.
Connection to Robust Optimization
In distributionally robust optimization (DRO):
where is an uncertainty set of distributions. When is convex and compact (in an appropriate topology) and the loss is convex in , the minimax theorem applies. This connects to adversarial training: the inner maximization finds the worst-case distribution, and the outer minimization hedges against it.
Canonical Examples
Matrix game (Rock-Paper-Scissors)
The payoff matrix for Rock-Paper-Scissors is:
The row player maximizes , the column player minimizes. Strategy spaces are the 3-simplex (compact, convex). The payoff is bilinear (convex-concave in mixed strategies). By Von Neumann's theorem, the game has value 0 and the unique optimal strategy for both players is .
Strong duality in LP
Primal: subject to , . Dual: subject to , , . The dual optimum is , giving value 12. The primal optimum is , also giving value 12. Strong duality holds: . The Lagrangian has a saddle point at .
Common Confusions
Minimax is not the same as min-max (sequential optimization)
In , the min-player commits to first, then the max-player responds optimally. This is different from alternating gradient descent-ascent, where both players update simultaneously. Alternating updates can cycle or diverge for non-convex-concave problems. The minimax theorem is about the existence of a saddle point, not about an algorithm to find it.
GANs do not satisfy the minimax theorem assumptions
The GAN objective is non-convex in the generator parameters. Von Neumann's theorem does not apply. The "theoretical optimum" analysis (showing the global optimum is ) assumes the discriminator is optimized exactly for each , which never happens in practice. GAN training is a non-convex saddle-point problem without convergence guarantees from classical minimax theory.
Weak duality always holds, strong duality requires conditions
is true for any (the "max-min inequality"). This is a one-line proof: , then take on the left and on the right. Equality (strong duality) requires structure: convexity-concavity plus compactness, or Slater's condition, or specific problem structure like LP.
Summary
- Weak duality: always holds
- Von Neumann: equality holds for continuous convex-concave on compact convex sets
- Saddle point :
- Strong LP duality is a corollary of the minimax theorem
- GANs are minimax problems but violate the convexity assumption
- Robust optimization: minimax over uncertainty sets
Exercises
Problem
Consider the function on , . Find and . Does a saddle point exist?
Problem
Prove the weak duality inequality for any function , assuming the max and min exist.
References
Canonical:
- Von Neumann, "Zur Theorie der Gesellschaftsspiele" (1928)
- Sion, "On General Minimax Theorems" (1958)
Current:
- Boyd & Vandenberghe, Convex Optimization (2004), Chapter 5 (duality)
- Goodfellow et al., "Generative Adversarial Nets" (2014)
- Ben-Tal, El Ghaoui, Nemirovski, Robust Optimization (2009), Chapters 1-2
Next Topics
- Adversarial machine learning: minimax formulations for robustness
- Robust optimization: worst-case optimization over uncertainty sets
Last reviewed: April 2026
Prerequisites
Foundations this topic depends on.
- Convex Optimization BasicsLayer 1
- Differentiation in RnLayer 0A
- Sets, Functions, and RelationsLayer 0A
- Basic Logic and Proof TechniquesLayer 0A
- Matrix Operations and PropertiesLayer 0A
- Convex DualityLayer 0B