Numerical Optimization
Interior Point Methods
Barrier functions transform constrained optimization into unconstrained problems. Newton steps on the barrier objective trace the central path to the constrained optimum with polynomial convergence.
Prerequisites
Why This Matters
Many optimization problems in ML and statistics involve constraints: positive-definiteness of covariance matrices (semidefinite programs), probability simplex constraints, box constraints on parameters. The simplex method for linear programming walks along the boundary of the feasible region and can take exponentially many steps in the worst case. Interior point methods take a different approach: they stay strictly inside the feasible region, using a barrier function to repel iterates from constraint boundaries. The result is polynomial-time convergence for LPs, quadratic programs, and semidefinite programs. Interior point methods are the standard solver backend for large-scale convex optimization.
Mental Model
Imagine trying to minimize a function inside a room without touching the walls. A barrier function adds a penalty that goes to infinity as you approach any wall. The penalty keeps you away from the boundaries. By gradually reducing the strength of the penalty, you allow the iterates to approach the boundary (where the constrained optimum typically lies) in a controlled manner. Newton's method on the penalized objective gives fast local convergence at each penalty level.
Formal Setup
Consider the constrained convex optimization problem:
where are convex and twice differentiable.
Logarithmic Barrier Function
The logarithmic barrier for the inequality constraints is:
This function is defined only for strictly feasible points (where for all ). It approaches as any constraint approaches equality. The gradient pushes the iterate away from constraint boundaries.
Barrier Problem
For a parameter , the barrier problem is:
As , the barrier term becomes negligible relative to , and the solution of the barrier problem approaches the solution of the original constrained problem.
Central Path
The central path is the curve of solutions where is the minimizer of . As increases, traces a smooth path from the analytic center of the feasible region (at ) toward the constrained optimum (as ).
The Barrier Method Algorithm
The barrier method repeatedly solves the barrier problem for increasing values of :
- Initialize with a strictly feasible point , parameter , and growth factor
- Centering step: starting from the current iterate, run Newton's method on until convergence to
- Update: set
- Stop when (duality gap bound)
Each centering step is an unconstrained minimization (Newton's method on a smooth, self-concordant function). The outer loop increases , tightening the approximation to the original problem.
Core Theory
Convergence of the Barrier Method
Statement
The barrier method produces a point with after at most:
outer iterations, where is the number of inequality constraints, is the initial barrier parameter, and is the growth factor. Each outer iteration requires Newton steps if (the theoretically optimal growth rate). The total number of Newton steps is .
Intuition
At any point on the central path with parameter , the suboptimality is at most . To reach suboptimality , we need . Starting from and multiplying by each iteration, we reach in iterations. Each centering step takes few Newton steps because is a warm start for (the central path is smooth).
Proof Sketch
The duality gap at is exactly (from the KKT conditions of the barrier problem: the dual variables are , and strong duality gives the gap). So . For the Newton step count: the barrier function is self-concordant (Nesterov and Nemirovski, 1994), and Newton's method on self-concordant functions converges in iterations from a warm start when is moderate. Choosing balances outer iterations against Newton steps per iteration.
Why It Matters
This gives polynomial-time convergence for any convex program with a self-concordant barrier. For linear programs with constraints and variables, the total complexity is (the comes from solving the Newton system). This is polynomial, unlike the simplex method which can take steps in worst-case examples (Klee-Minty).
Failure Mode
The method requires a strictly feasible starting point. Finding one can be as hard as solving the original problem (Phase I methods address this). The Newton system at each step requires solving an linear system, which costs in general. For very large , this is the bottleneck, and iterative solvers or exploiting problem structure is necessary. Also, the barrier function is not defined at the boundary, so the method cannot handle equality constraints directly (they must be eliminated or handled separately).
Why Interior Point Methods Dominate
Linear programming. The simplex method is fast in practice (polynomial on average) but exponential in the worst case. Interior point methods guarantee iterations. For large, sparse LPs, interior point methods are competitive with or faster than simplex.
Semidefinite programming (SDP). There is no analog of the simplex method for SDPs. Interior point methods, using the log-determinant barrier for positive semidefinite matrix constraints, are the primary solution method. Many ML problems (kernel learning, robust PCA, relaxations of combinatorial problems) reduce to SDPs.
Second-order cone programming (SOCP). Interior point methods handle second-order cone constraints with the same framework, using appropriate self-concordant barriers.
Connection to KKT Conditions
The centering step computes a point where:
Defining , this becomes with and . As , , approaching complementary slackness. The barrier method traces a smooth path toward the KKT conditions.
Common Confusions
The barrier parameter t is not a step size
The parameter controls the fidelity of the barrier approximation, not the Newton step size. Increasing makes the barrier problem closer to the original constrained problem. Newton's method has its own step size (determined by line search) within each centering step.
Interior point methods do not find vertices
For linear programs, the simplex method finds a vertex (extreme point) of the feasible polyhedron. Interior point methods find a point in the interior, near but not exactly at a vertex. If an exact vertex solution is needed (e.g., for integer programming), a crossover procedure can be applied to the interior point solution.
Key Takeaways
- The log barrier keeps iterates strictly feasible while penalizing proximity to boundaries
- The central path traces a smooth curve from the analytic center to the constrained optimum as
- Each centering step uses Newton's method on the barrier objective
- Total complexity is Newton steps for constraints
- Interior point methods provide polynomial convergence for LPs, SDPs, and SOCPs
- A strictly feasible starting point is required; Phase I methods find one if it exists
Exercises
Problem
Consider the 1D problem: minimize subject to and . Write the barrier function . For , find the minimizer of the barrier problem .
Problem
Show that for a linear program with constraints, the duality gap at the point on the central path is exactly . Derive the dual variables from the barrier optimality conditions.
References
Canonical:
- Boyd and Vandenberghe, Convex Optimization (2004), Chapter 11
- Nesterov and Nemirovski, Interior-Point Polynomial Algorithms in Convex Programming (1994)
Current:
- Nocedal and Wright, Numerical Optimization (2006), Chapter 14
- Wright, Primal-Dual Interior-Point Methods (1997)
Next Topics
- Primal-dual interior point methods and Mehrotra predictor-corrector
- Semidefinite programming and applications in ML
Last reviewed: April 2026
Prerequisites
Foundations this topic depends on.
- Convex Optimization BasicsLayer 1
- Differentiation in RnLayer 0A
- Sets, Functions, and RelationsLayer 0A
- Basic Logic and Proof TechniquesLayer 0A
- Matrix Operations and PropertiesLayer 0A
- Newton's MethodLayer 1