Learning Track
Engineer to ML Theory
You can train models and ship code. This path helps you understand why things work, when they break, and what the theorems actually say. No unnecessary math. Just the theory that changes how you think about models.
Phase 1: Why Your Model Works (or Doesn't)
You already use these tools. Now understand what they're doing mathematically.
Phase 2: The Optimizer You're Actually Using
Adam, SGD, learning rate schedules. What the math says about convergence.
Phase 3: Transformers from the Inside
Not just how to use them. How they compute, what the math looks like, and why they work.
Phase 4: Scaling and Training at Scale
The math behind training large models. Scaling laws, RLHF, and post-training.
Phase 5: Why Things Break
The theory that explains failure modes. Double descent, grokking, interpretability.
Already know some theory?
Skip to whatever phase matches your current understanding. Each page shows prerequisites in the sidebar so you can always check what you might be missing.
Or ask a specific question →