Pages that link to "Item:Q4594841"
From MaRDI portal
The following pages link to Non-asymptotic convergence analysis of inexact gradient methods for machine learning without strong convexity (Q4594841):
Displaying 15 items.
- Asymptotic estimates for \(r\)-Whitney numbers of the second kind (Q1714572) (← links)
- Incremental without replacement sampling in nonconvex optimization (Q2046568) (← links)
- From inexact optimization to learning via gradient concentration (Q2111477) (← links)
- A frequency-domain analysis of inexact gradient methods (Q2149575) (← links)
- Accelerating incremental gradient optimization with curvature information (Q2181597) (← links)
- A linearly convergent stochastic recursive gradient method for convex optimization (Q2228399) (← links)
- Quadratic optimization with orthogonality constraint: explicit Łojasiewicz exponent and linear convergence of retraction-based line-search and stochastic variance-reduced gradient methods (Q2330648) (← links)
- Restricted strong convexity and its applications to convergence analysis of gradient-type methods in convex optimization (Q2355319) (← links)
- Inexact gradient projection method with relative error tolerance (Q2696906) (← links)
- RSG: Beating Subgradient Method without Smoothness and Strong Convexity (Q4558142) (← links)
- On the Estimation Performance and Convergence Rate of the Generalized Power Method for Phase Synchronization (Q4602339) (← links)
- On the Nonergodic Convergence Rate of an Inexact Augmented Lagrangian Framework for Composite Convex Programming (Q5219732) (← links)
- Convergence Analysis of Inexact Randomized Iterative Methods (Q5856678) (← links)
- Inexact SARAH algorithm for stochastic optimization (Q5859016) (← links)
- Hölderian Error Bounds and Kurdyka-Łojasiewicz Inequality for the Trust Region Subproblem (Q5870365) (← links)