Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates
From MaRDI portal
Publication:6330368
arXiv1912.01597MaRDI QIDQ6330368
Peter Richtárik, Konstantin Mishchenko, Dmitry Kovalev
Publication date: 3 December 2019
Abstract: We present two new remarkably simple stochastic second-order methods for minimizing the average of a very large number of sufficiently smooth and strongly convex functions. The first is a stochastic variant of Newton's method (SN), and the second is a stochastic variant of cubically regularized Newton's method (SCN). We establish local linear-quadratic convergence results. Unlike existing stochastic variants of second order methods, which require the evaluation of a large number of gradients and/or Hessians in each iteration to guarantee convergence, our methods do not have this shortcoming. For instance, the simplest variants of our methods in each iteration need to compute the gradient and Hessian of a {em single} randomly selected function only. In contrast to most existing stochastic Newton and quasi-Newton methods, our approach guarantees local convergence faster than with first-order oracle and adapts to the problem's curvature. Interestingly, our method is not unbiased, so our theory provides new intuition for designing new stochastic methods.
Has companion code repository: https://github.com/nathansiae/Stochastic-Average-Newton
This page was built for publication: Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6330368)