Efficient Second-order Optimization for Machine Learning

3 764
22.8
Опубликовано 8 февраля 2018, 23:32
Stochastic gradient-based methods are the state-of-the-art in large-scale machine learning optimization due to their extremely efficient per-iteration computational cost. Second-order methods, that use the second derivative of the optimization objective, are known to enable faster convergence. However, the latter has been much less explored due to the high cost of computing the second-order information. We will present second-order stochastic methods for (convex and non-convex) optimization problems arising in machine learning that match the per-iteration cost of gradient-based methods, yet enjoy the faster convergence properties of second-order optimization overall leading to faster algorithms than the best known gradient-based methods.

See more at microsoft.com/en-us/research/v...
Свежие видео
1 день – 9 9306:10
Intro to AI agents
9 дней – 2 3901:10
AGM H MAX|Enjoy More Charge Less
14 дней – 86 2426:26
I'm giving away 14 years of drawings
автотехномузыкадетское