Best subset, forward stepwise or Lasso? Analysis and recommendations based on extensive comparisons
From MaRDI portal
Publication:2225312
DOI10.1214/19-STS733OpenAlexW4210643453MaRDI QIDQ2225312
Robert Tibshirani, Ryan J. Tibshirani, Trevor Hastie
Publication date: 8 February 2021
Published in: Statistical Science (Search for Journal in Brave)
Full work available at URL: https://projecteuclid.org/euclid.ss/1605603631
Related Items
A literature review of (Sparse) exponential family PCA, Variable Selection With Second-Generation P-Values, Clustering, multicollinearity, and singular vectors, An Alternating Method for Cardinality-Constrained Optimization: A Computational Study for the Best Subset Selection and Sparse Portfolio Problems, Predicting the equity market risk premium: a model selection approach, Robust grouped variable selection using distributionally robust optimization, Stacked grenander and rearrangement estimators of a discrete distribution, Simultaneous feature selection and outlier detection with optimality guarantees, Subset Selection for Linear Mixed Models, Variables selection using \(\mathcal{L}_0\) penalty, Comparing Bayesian variable selection to Lasso approaches for applications in psychology, Robust variable selection and estimation via adaptive elastic net S-estimators for linear regression, Nonlinear Iterative Hard Thresholding for Inverse Scattering, An automated exact solution framework towards solving the logistic regression best subset selection problem, Optimal learning, A unified precision matrix estimation framework via sparse column-wise inverse operator under weak sparsity, Distributed primal outer approximation algorithm for sparse convex programming with separable structures, Subset Selection and the Cone of Factor-Width-k Matrices, Unnamed Item, Unnamed Item, Sparse regression: scalable algorithms and empirical performance, Rejoinder: ``Sparse regression: scalable algorithms and empirical performance, Iteratively reweighted \(\ell_1\)-penalized robust regression, An offline/online procedure for dual norm calculations of parameterized functionals: empirical quadrature and empirical test spaces, Sparse classification: a scalable discrete optimization perspective, Robust subset selection, New bounds for subset selection from conic relaxations, Dismemberment and design for controlling the replication variance of regret for the multi-armed bandit, Variable selection in convex quantile regression: \(\mathcal{L}_1\)-norm or \(\mathcal{L}_0\)-norm regularization?, Unnamed Item, Unnamed Item, Sparse regression at scale: branch-and-bound rooted in first-order optimization
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Nearly unbiased variable selection under minimax concave penalty
- Best subset selection via a modern optimization lens
- Degrees of freedom in lasso problems
- Improved variable selection with forward-lasso adaptive shrinkage
- Relaxed Lasso
- Least angle regression. (With discussion)
- The Lasso problem and uniqueness
- Sparse high-dimensional regression: exact scalable algorithms and phase transitions
- Pathwise coordinate optimization
- On the ``degrees of freedom of the lasso
- The Dantzig selector: statistical estimation when \(p\) is much larger than \(n\). (With discussions and rejoinder).
- When does more regularization imply fewer degrees of freedom? Sufficient conditions and counterexamples
- Square-root lasso: pivotal recovery of sparse signals via conic programming
- SparseNet: Coordinate Descent With Nonconvex Penalties
- Regressions by Leaps and Bounds
- Atomic Decomposition by Basis Pursuit
- Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties
- Sparse Approximate Solutions to Linear Systems
- Matching pursuits with time-frequency dictionaries
- Effective degrees of freedom: a flawed metaphor
- Strong Rules for Discarding Predictors in Lasso-Type Problems