Accelerating Stochastic Gradient Descent

8 925
23.2
Опубликовано 24 июля 2017, 6:00
There is widespread sentiment that it is not possible to effectively utilize fast gradient methods (e.g. Nesterov's acceleration, conjugate gradient, heavy ball) for the purposes of stochastic optimization due to their instability and error accumulation, a notion made precise in dAspremont 2008 and Devolder, Glineur, and Nesterov 2014. This work considers the use of "fast gradient" methods for the special case of stochastic approximation for the least squares regression problem. Our main result refutes the conventional wisdom by showing that acceleration can be made robust to statistical errors. In particular, this work introduces an accelerated stochastic gradient method that provably achieves the minimax optimal statistical risk faster than stochastic gradient descent. Critical to the analysis is a sharp characterization of accelerated stochastic gradient descent as a stochastic process.

See more on this video at microsoft.com/en-us/research/v...
Случайные видео
364 дня – 802 7000:23
This PC Setup has 92 Screens...
05.10.21 – 644 5644:47
Why Do Command Lines Still Exist?
07.02.08 – 110 4855:43
What's In My Dock
автотехномузыкадетское