The following pages link to (Q5149016):
Displaying 13 items.
- Incremental without replacement sampling in nonconvex optimization (Q2046568) (← links)
- An adaptive Polyak heavy-ball method (Q2102380) (← links)
- Machine learning design of volume of fluid schemes for compressible flows (Q2123342) (← links)
- Random Batch Methods for Classical and Quantum Interacting Particle Systems and Statistical Samplings (Q5054578) (← links)
- Stochastic momentum methods for non-convex learning without bounded assumptions (Q6057975) (← links)
- SVRG meets AdaGrad: painless variance reduction (Q6097116) (← links)
- Stochastic Gauss-Newton algorithms for online PCA (Q6111665) (← links)
- Convergence Properties of an Objective-Function-Free Optimization Regularization Algorithm, Including an \(\boldsymbol{\mathcal{O}(\epsilon^{-3/2})}\) Complexity Bound (Q6116246) (← links)
- Adaptive step size rules for stochastic optimization in large-scale learning (Q6116586) (← links)
- An adaptive Riemannian gradient method without function evaluations (Q6167088) (← links)
- Gradient descent in the absence of global Lipschitz continuity of the gradients (Q6583712) (← links)
- An improved transformer model with multi-head attention and attention to attention for low-carbon multi-depot vehicle routing problem (Q6589088) (← links)
- High probability bounds on AdaGrad for constrained weakly convex optimization (Q6649705) (← links)