Optimal detection of sparse principal components in high dimension

From MaRDI portal
Publication:385763

DOI10.1214/13-AOS1127zbMath1277.62155arXiv1202.5070OpenAlexW2006452405WikidataQ59409960 ScholiaQ59409960MaRDI QIDQ385763

Quentin Berthet, Philippe Rigollet

Publication date: 11 December 2013

Published in: The Annals of Statistics (Search for Journal in Brave)

Full work available at URL: https://arxiv.org/abs/1202.5070



Related Items

Isotonic regression with unknown permutations: statistics, computation and adaptation, Tensor clustering with planted structures: statistical optimality and computational limits, Wald Statistics in high-dimensional PCA, Exploring dimension learning via a penalized probabilistic principal component analysis, Hypothesis testing for high-dimensional multinomials: a selective review, Sharp minimax tests for large covariance matrices and adaptation, Optimal multiple change-point detection for high-dimensional data, Multidimensional two-component Gaussian mixtures detection, Gaussian determinantal processes: A new model for directionality in data, Recent developments in high dimensional covariance estimation and its related issues, a review, Bayesian inference for spectral projectors of the covariance matrix, Testing the order of a population spectral distribution for high-dimensional data, High-dimensional change-point estimation: combining filtering with convex optimization, On the optimality of sliced inverse regression in high dimensions, Asymptotic power of sphericity tests for high-dimensional data, Testing for principal component directions under weak identifiability, Efficient estimation of linear functionals of principal components, Using ℓ1-Relaxation and Integer Programming to Obtain Dual Bounds for Sparse PCA, Large covariance estimation through elliptical factor models, Minimax estimation in sparse canonical correlation analysis, Solving sparse principal component analysis with global support, Sharp detection boundaries on testing dense subhypergraph, Matrix means and a novel high-dimensional shrinkage phenomenon, Estimation of Wasserstein distances in the spiked transport model, Estimation of functionals of sparse covariance matrices, Unnamed Item, Free Energy Wells and Overlap Gap Property in Sparse PCA, Sparse signal reconstruction via the approximations of \(\ell_0\) quasinorm, Community detection in sparse random networks, Guaranteed recovery of planted cliques and dense subgraphs by convex relaxation, Finding hidden cliques of size \(\sqrt{N/e}\) in nearly linear time, A wonderful triangle in compressed sensing, Sparse principal component analysis for high‐dimensional stationary time series, Fundamental limits of detection in the spiked Wigner model, Statistical and computational limits for sparse matrix detection, Public-key encryption from homogeneous CLWE, Inference for low-rank models, On lower bounds for the bias-variance trade-off, Phase transitions for detecting latent geometry in random graphs, Sparse PCA: optimal rates and adaptive estimation, Optimal rates of statistical seriation, The spectral norm of random inner-product kernel matrices, Notes on computational-to-statistical gaps: predictions using statistical physics, Combinatorial inference for graphical models, High-dimensional covariance matrices in elliptical distributions with application to spherical test, Approximation bounds for sparse principal component analysis, Comment on ``Hypothesis testing by convex optimization, Robust covariance estimation for approximate factor models, The limits of the sample spiked eigenvalues for a high-dimensional generalized Fisher matrix and its applications, Unnamed Item, Slope meets Lasso: improved oracle bounds and optimality, Finding a large submatrix of a Gaussian random matrix, Detecting Markov random fields hidden in white noise, An $\ell_{\infty}$ Eigenvector Perturbation Bound and Its Application to Robust Covariance Estimation, Community Detection and Stochastic Block Models, Projection tests for high-dimensional spiked covariance matrices, Tests for covariance matrices in high dimension with less sample size, Sparse power factorization: balancing peakiness and sample complexity, Optimal testing for planted satisfiability problems, Proximal Distance Algorithms: Theory and Examples, ECA: High-Dimensional Elliptical Component Analysis in Non-Gaussian Distributions, Two-sample Hypothesis Testing for Inhomogeneous Random Graphs, Community detection in dense random networks, Finding a planted clique by adaptive probing, Sparse equisigned PCA: algorithms and performance bounds in the noisy rank-1 setting, Recovery of simultaneous low rank and two-way sparse coefficient matrices, a nonconvex approach, Estimating structured high-dimensional covariance and precision matrices: optimal rates and adaptive estimation, Minimax rates in sparse, high-dimensional change point detection, Optimality and sub-optimality of PCA. I: Spiked random matrix models, Sequential subspace change point detection, A robust test for sphericity of high-dimensional covariance matrices, Scale-Invariant Sparse PCA on High-Dimensional Meta-Elliptical Data, Hypothesis testing for densities and high-dimensional multinomials: sharp local minimax rates, Sparse principal component analysis with missing observations, Algorithmic thresholds for tensor PCA, Sparsistency and agnostic inference in sparse PCA, Optimal estimation and rank detection for sparse spiked covariance matrices, Notes on computational hardness of hypothesis testing: predictions using the low-degree likelihood ratio, Detecting positive correlations in a multivariate sample, Computational barriers in minimax submatrix detection, Do semidefinite relaxations solve sparse PCA up to the information limit?



Cites Work