Outline
The 2018 course consists of the following topics
lecture 1 | Introduction to convex optimization and iterative methods. |
lecture 2 | Review of basic probability theory. |
Maximum likelihood, M-estimators, and empirical risk minimization as a motivation for convex optimization. | |
lecture 3 | Fundamental concepts in convex analysis. |
Basics of complexity theory. | |
lecture 4 | Unconstrained smooth minimization I: |
Concept of an iterative optimization algorithm. | |
Convergence rates. | |
Characterization of functions. | |
lecture 5 |
Unconstrained smooth minimization II:
|
Gradient and accelerated gradient methods. | |
lecture 6 | Unconstrained smooth minimization III: |
The quadratic case. | |
The conjugate gradient method. | |
Variable metric algorithms. | |
lecture 7 | Stochastic gradient methods. |
lecture 8 | Non-convex optimization. |
Neural networks. | |
Convergence of SGD on nonconvex problems. | |
lecture 9 | Composite convex minimization I. |
Subgradient method. | |
Proximal and accelerated proximal gradient methods. | |
lecture 10 | Composite convex minimization II. |
Proximal Newton-type methods. | |
Stochastic proximal gradient methods. | |
lecture 11 | Constrained convex minimization I. |
The primal-dual approach. | |
Smoothing approaches for non-smooth convex minimization. | |
lecture 12 | Constrained convex minimization II. |
The Frank-Wolfe method. | |
The universal primal-dual gradient method. | |
The alternating direction method of multipliers (ADMM). |