Outline
The 2017 course consists of the following topics
| Lecture 1 | Introduction to convex optimization and iterative methods. |
| Lecture 2 | Review of basic probability theory. |
| Maximum likelihood, M-estimators, and empirical risk minimization as a motivation for convex optimization. | |
| Lecture 3 | Fundamental concepts in convex analysis. |
| Basics of complexity theory. | |
| Lecture 4 | Unconstrained smooth minimization I: |
| Concept of an iterative optimization algorithm. | |
| Convergence rates. | |
| Characterization of functions. | |
| Lecture 5 |
Unconstrained smooth minimization II:
|
| Gradient and accelerated gradient methods. | |
| Lecture 6 | Unconstrained smooth minimization III: |
| The quadratic case. | |
| The conjugate gradient method. | |
| Variable metric algorithms. | |
| Lecture 7 | Stochastic gradient methods. |
| Lecture 8 | Composite convex minimization I: |
| Subgradient method. | |
| Proximal and accelerated proximal gradient methods. | |
| Lecture 9 | Composite convex minimization II: |
| Proximal Newton-type methods. | |
| Stochastic proximal gradient methods. | |
| Lecture 10 | Constrained convex minimization I: |
| The primal-dual approach. | |
| Smoothing approaches for non-smooth convex minimization. | |
| Lecture 11 | Constrained convex minimization II: |
| The Frank-Wolfe method. | |
| The universal primal-dual gradient method. | |
| The alternating direction method of multipliers (ADMM). | |
| Lecture 12 | Disciplined convex programming. |