Applied ML
Lyapunov-Based Machine Learning for Chaos
Physics-informed losses that match Lyapunov spectra, Hamiltonian and Lagrangian neural networks, neural ODEs that respect known invariants, and the Lyapunov-time wall on prediction horizons.
Why This Matters
Standard MSE training on chaotic trajectory data fails for a structural reason. Pointwise prediction loss collapses past one Lyapunov time: any model that nails the next accumulates exponential error after that, and the gradient signal becomes uninformative about the long-term geometry of the attractor. The model can be Pareto-optimal on short-horizon MSE while producing trajectories that drift off the attractor, fail energy conservation, or settle into spurious fixed points.
The fix is to train on geometric and dynamical invariants, not just trajectory samples. Match the Lyapunov spectrum. Enforce symplectic structure where the true system is Hamiltonian. Constrain the learned vector field to known conservation laws. The result is a model that may have worse one-step error but stays on the right manifold for arbitrary horizons.
This matters wherever ML meets physical dynamics: weather emulators, plasma simulators, molecular dynamics surrogates, climate downscaling, fusion control. The Lyapunov-time wall is real; the engineering question is what you ask the model to preserve once you cross it.
Core Ideas
Hamiltonian neural networks (HNN). Greydanus, Dzamba, and Yosinski (NeurIPS 2019; arXiv 1906.01563) parameterize a scalar and define the vector field by Hamilton's equations , . Training matches observed time derivatives to those induced by . Energy is conserved by construction up to integrator error, even off the training distribution. The cost is the architectural commitment: HNNs only apply where a Hamiltonian formulation exists.
Lagrangian neural networks (LNN). Cranmer et al. (ICLR 2020 workshop; arXiv 2003.04630) parameterize a Lagrangian and recover dynamics through the Euler-Lagrange equations, which require inverting the mass matrix . LNNs handle systems with non-Cartesian coordinates and constraints that HNNs struggle with, at the cost of a Hessian inversion in the forward pass.
Lyapunov-spectrum matching. Add a loss term penalizing the difference between the maximal Lyapunov exponent of the learned model and the value estimated from data via the Wolf or Rosenstein algorithms. This is more expensive than trajectory MSE: each gradient step requires propagating a tangent vector or solving a variational equation along a sampled trajectory. The payoff is that the learned attractor has the right stretching rate, so short-horizon predictions plus invariant statistics both match.
Reservoir computing for chaos. Pathak, Hunt, Girvan, Lu, and Ott (Physical Review Letters 120, 2018; arXiv 1710.07313) showed that echo-state networks can predict chaotic flows several Lyapunov times into the future and, more importantly, reproduce the climatology and Lyapunov spectrum of the true system. This established the empirical baseline that more elaborate architectures are measured against. The result is sharp: prediction skill extends to roughly to for Kuramoto-Sivashinsky, then collapses as expected.
The Lyapunov-time wall. Any deterministic model, regardless of capacity, loses meaningful pointwise predictive skill after a few Lyapunov times. Initial condition uncertainty saturates the attractor diameter at time . Once you hit the wall, the correct objective shifts to ensemble forecasts, distributional metrics, and invariant statistics. Pretending the wall does not exist is the most common failure mode of ML weather and climate work.
Common Confusions
Energy conservation does not imply correct dynamics
A neural network that conserves energy exactly can still produce trajectories that miss the attractor or settle into the wrong invariant manifold. Symplectic structure is necessary but not sufficient. Validation needs Lyapunov spectra, power spectra, and invariant measure comparisons, not just energy drift plots.
Short-horizon MSE overstates skill on chaotic systems
A model with one-step MSE smaller than baseline can produce climatologies that look nothing like the true system. Short-horizon point error is the wrong metric past . Report invariant statistics: marginal distributions, autocorrelation, Lyapunov exponents, and attractor reconstruction quality.
References
Related Topics
Last reviewed: April 18, 2026
Prerequisites
Foundations this topic depends on.
- Nonlinear Dynamics and Chaos FundamentalsLayer 4
- Stochastic Differential EquationsLayer 3
- Brownian MotionLayer 2
- Measure-Theoretic ProbabilityLayer 0B
- Martingale TheoryLayer 0B
- Ito's LemmaLayer 3
- Stochastic Calculus for MLLayer 3
- Fokker–Planck EquationLayer 3
- PDE Fundamentals for Machine LearningLayer 1
- Fast Fourier TransformLayer 1
- Exponential Function PropertiesLayer 0A
- Eigenvalues and EigenvectorsLayer 0A
- Matrix Operations and PropertiesLayer 0A
- Sets, Functions, and RelationsLayer 0A
- Basic Logic and Proof TechniquesLayer 0A
- Functional Analysis CoreLayer 0B
- Metric Spaces, Convergence, and CompletenessLayer 0A
- Inner Product Spaces and OrthogonalityLayer 0A
- Vectors, Matrices, and Linear MapsLayer 0A
- Physics-Informed Neural NetworksLayer 4
- The Jacobian MatrixLayer 0A
- Automatic DifferentiationLayer 1
- Feedforward Networks and BackpropagationLayer 2
- Differentiation in RnLayer 0A
- Matrix CalculusLayer 1
- The Hessian MatrixLayer 0A
- Activation FunctionsLayer 1
- Convex Optimization BasicsLayer 1
- Gradient Descent VariantsLayer 1