Publication | Closed Access
Note on Learning Rate Schedules for Stochastic Optimization
133
Citations
5
References
1990
Year
Mathematical ProgrammingModel OptimizationEngineeringMachine LearningData ScienceStochastic OptimizationPattern RecognitionGeneral AlgorithmOnline AlgorithmStochastic Gradient DescentComputational Learning TheoryRate SchedulesLarge Scale OptimizationComputer ScienceDeep LearningAdaptive Optimization
We present and compare learning rate schedules for stochastic gradient descent, a general algorithm which includes LMS, on-line backpropagation and k-means clustering as special cases. We introduce search-then-converge type schedules which outperform the classical constant and running average (1/t) schedules both in speed of convergence and quality of solution.
| Year | Citations | |
|---|---|---|
Page 1
Page 1