A second-order method for strongly convex \(\ell _1\)-regularization problems
From MaRDI portal
Publication:263191
DOI10.1007/s10107-015-0875-4zbMath1364.90255arXiv1306.5386OpenAlexW2164449950MaRDI QIDQ263191
Kimon Fountoulakis, Jacek Gondzio
Publication date: 4 April 2016
Published in: Mathematical Programming. Series A. Series B (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1306.5386
\(\ell_1\)-regularizationiteration complexityNewton conjugate-gradients methodsecond-order methodsstrongly convex optimization
Analysis of algorithms (68W40) Numerical mathematical programming methods (65K05) Convex programming (90C25) Large-scale problems in mathematical programming (90C06)
Related Items
Mathematical optimization in classification and regression trees, Sparse Approximations with Interior Point Methods, Performance of first- and second-order methods for \(\ell_1\)-regularized least squares problems, A Preconditioner for A Primal-Dual Newton Conjugate Gradient Method for Compressed Sensing Problems, A flexible coordinate descent method, Visualizing proportions and dissimilarities by space-filling maps: a large neighborhood search approach, On the convergence rate of scaled gradient projection method, An active set Newton-CG method for \(\ell_1\) optimization, Optimal randomized classification trees, Gradient-based method with active set strategy for $\ell _1$ optimization, Continuation methods for approximate large scale object sequencing, On sparse ensemble methods: an application to short-term predictions of the evolution of COVID-19, Visualizing data as objects by DC (difference of convex) optimization, Second-order orthant-based methods with enriched Hessian information for sparse \(\ell _1\)-optimization, A Fast Active Set Block Coordinate Descent Algorithm for $\ell_1$-Regularized Least Squares, Interior-point solver for convex separable block-angular problems, Generalized Conjugate Gradient Methods for ℓ1 Regularized Convex Quadratic Programming with Finite Convergence, On partial Cholesky factorization and a variant of quasi-Newton preconditioners for symmetric positive definite matrices, Linesearch Newton-CG methods for convex optimization with noise, An inexact dual logarithmic barrier method for solving sparse semidefinite programs
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems
- Parallel coordinate descent methods for big data optimization
- A coordinate gradient descent method for nonsmooth separable minimization
- CoSaMP: Iterative signal recovery from incomplete and inaccurate samples
- Templates for convex cone problems with applications to sparse signal recovery
- Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function
- Coordinate descent algorithms for lasso penalized regression
- A Mathematical View of Interior-Point Methods in Convex Optimization
- Accelerated Block-coordinate Relaxation for Regularized Optimization
- Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems
- NESTA: A Fast and Accurate First-Order Method for Sparse Recovery
- A Nonlinear Primal-Dual Method for Total Variation-Based Image Restoration
- Analysis of bounded variation penalty methods for ill-posed problems
- Multiple View Geometry in Computer Vision
- Convergence of a block coordinate descent method for nondifferentiable minimization