An Introduction To Optimization Techniques In Machine Learning

- From linear programming to non-convex smooth optimization
 - Gradient descent
 - Polyak's Heavy ball: accelerated gradient descent
 - Stochastic gradient descent (SGD)
 - SGD and Markov chains
 - Variance reduced SGD
 - Computational efficiency of SGD for statistical inference
 - Computational statistical tradeoff through convex relaxation
 
Modules
