Efficient Second-order Optimization for Machine Learning

3 745
22.7
Следующее
09.02.18 – 1 7261:07:07
Numerical Tools for Non-Experts
Популярные
Опубликовано 8 февраля 2018, 23:32
Stochastic gradient-based methods are the state-of-the-art in large-scale machine learning optimization due to their extremely efficient per-iteration computational cost. Second-order methods, that use the second derivative of the optimization objective, are known to enable faster convergence. However, the latter has been much less explored due to the high cost of computing the second-order information. We will present second-order stochastic methods for (convex and non-convex) optimization problems arising in machine learning that match the per-iteration cost of gradient-based methods, yet enjoy the faster convergence properties of second-order optimization overall leading to faster algorithms than the best known gradient-based methods.

See more at microsoft.com/en-us/research/v...
автотехномузыкадетское