@lyc102
2018-01-26T12:36:10.000000Z
字数 2530
阅读 1986
machine_learning
Consder the minimization problem
GD methods:
Theorem. If is 1) first order Lip continuous and 2) stricktly convex, then with , there exists s.t.
Pro: fast and easy to implement.
Disadvantanges:
- Hard to chose
- might be a vector not a scalar
- stuck in a local minimum or saddle points
Define Regret
Full Gradient
SGD
Consider the least squares problem
Apply SGD to solve the least squares problem and present convergence proof.