Collection-landing

Online Learning Crash Course

A foundational journey into sequential decision-making, regret minimization, and adaptive optimization algorithms.

This crash course provides a self-contained introduction to the fundamental principles and algorithms of online learning. We explore how to make optimal sequential decisions in the face of uncertainty, quantify performance using regret, and develop algorithms like Online Gradient Descent, Follow-The-Regularized-Leader, Mirror Descent, and adaptive methods such as AdaGrad.

The series emphasizes the mathematical underpinnings and geometric interpretations, laying the groundwork for understanding advanced optimization techniques in machine learning. Key topics include:

  • The Online Learning Protocol & Motivation
  • Regret Analysis & Performance Benchmarks
  • Gradient-Based Methods (OGD)
  • The Power of Regularization (FTRL)
  • Geometric Generalizations (Mirror Descent)
  • Adaptive Learning Rates (AdaGrad & Beyond)
  • Online-to-Batch Conversions & Generalization

This course is designed for readers comfortable with linear algebra, multivariable calculus, and basic mathematical notation. It serves as an essential prerequisite for deeper dives into optimization theory within machine learning, particularly for understanding adaptive optimizers and their connections to online methods.