Gradient Methods
Gradient descent, SGD, learning rate effects, convergence, convex vs non-convex, and saddle points.
0/59 done
Gradient Methods
7 lessons · Read before practicing
Gradient descent
- 1Gradient Descent Update RuleEasy
- 2Role of Learning RateEasy
- 3Batch Gradient Descent DisadvantageEasy
- 4Gradient Descent on Convex FunctionsMed
- 5Vanishing Gradient ProblemMed
- 6Momentum in Gradient DescentMed
- 7Gradient and Steepest DescentMed
- 8Gradient Descent OscillationMed
- 9Second-Order vs First-Order OptimizationHard
- 10Gradient ClippingHard