Rejoinder: ``A significance test for the lasso
From MaRDI portal
Publication:2249839
DOI10.1214/14-AOS1175REJzbMath1305.62255arXiv1405.6805MaRDI QIDQ2249839
Jonathan E. Taylor, Richard A. Lockhart, Ryan J. Tibshirani, Robert Tibshirani
Publication date: 3 July 2014
Published in: The Annals of Statistics (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1405.6805
Ridge regression; shrinkage estimators (Lasso) (62J07) Linear regression; mixed models (62J05) Parametric hypothesis testing (62F03) Generalized linear models (logistic models) (62J12)
Related Items
Testing Shape Constraints in Lasso Regularized Joinpoint Regression, Inference in adaptive regression via the Kac-Rice formula, Exact post-selection inference, with application to the Lasso, The use of vector bootstrapping to improve variable selection precision in Lasso models, The benefit of group sparsity in group inference with de-biased scaled group Lasso, AIC for the Lasso in generalized linear models, Scalable methods for Bayesian selective inference, The geometry of least squares in the 21st century, Controlling the false discovery rate via knockoffs, Targeted Inference Involving High-Dimensional Data Using Nuisance Penalized Regression, High-dimensional inference in misspecified linear models, Debiasing the Lasso: optimal sample size for Gaussian designs, MDR method for nonbinary response variable, Testing for high-dimensional network parameters in auto-regressive models, Linear hypothesis testing for high dimensional generalized linear models
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers
- A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems
- Inference in adaptive regression via the Kac-Rice formula
- Statistical significance in high-dimensional linear models
- Degrees of freedom in lasso problems
- Near-ideal model selection by \(\ell _{1}\) minimization
- High-dimensional variable selection
- Templates for convex cone problems with applications to sparse signal recovery
- Persistene in high-dimensional linear predictor-selection and the virtue of overparametrization
- Least angle regression. (With discussion)
- The Lasso problem and uniqueness
- Pathwise coordinate optimization
- On the ``degrees of freedom of the lasso
- Validity of the expected Euler characteristic heuristic
- Confidence Intervals and Hypothesis Testing for High-Dimensional Regression
- Hypothesis Testing in High-Dimensional Regression Under the Gaussian Random Design Model: Asymptotic Theory
- p-Values for High-Dimensional Regression
- NESTA: A Fast and Accurate First-Order Method for Sparse Recovery
- Scaled sparse linear regression
- A Perturbation Method for Inference on Regularized Regression Estimates
- Recovery of Exact Sparse Representations in the Presence of Bounded Noise
- Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies?
- How Biased is the Apparent Error Rate of a Prediction Rule?
- Estimation of Parameters and Larger Quantiles Based on the k Largest Observations
- Atomic Decomposition by Basis Pursuit
- A new approach to variable selection in least squares problems
- Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using $\ell _{1}$-Constrained Quadratic Programming (Lasso)
- Regularization and Variable Selection Via the Elastic Net
- Compressed sensing
- Discussion of: ``Grouping strategies and thresholding for high dimension linear models