Just interpolate: kernel ``ridgeless regression can generalize

From MaRDI portal
Publication:2196223

DOI10.1214/19-AOS1849zbMath1453.68155arXiv1808.00387OpenAlexW3104969455MaRDI QIDQ2196223

Tengyuan Liang, Alexander Rakhlin

Publication date: 28 August 2020

Published in: The Annals of Statistics (Search for Journal in Brave)

Full work available at URL: https://arxiv.org/abs/1808.00387




Related Items (39)

Canonical thresholding for nonsparse high-dimensional linear regressionMehler’s Formula, Branching Process, and Compositional Kernels of Deep Neural NetworksDeep Neural Networks, Generic Universal Interpolation, and Controlled ODEsDeep learning: a statistical viewpointFit without fear: remarkable mathematical phenomena of deep learning through the prism of interpolationCommunication-efficient distributed estimator for generalized linear models with a diverging number of covariatesSurprises in high-dimensional ridgeless least squares interpolationGeneralization error of random feature and kernel methods: hypercontractivity and kernel matrix concentrationOn the proliferation of support vectors in high dimensions*Learning from non-random data in Hilbert spaces: an optimal recovery perspectiveA precise high-dimensional asymptotic theory for boosting and minimum-\(\ell_1\)-norm interpolated classifiersTheoretical issues in deep networksBenign overfitting in linear regressionOverparameterization and Generalization Error: Weighted Trigonometric InterpolationImproved complexities for stochastic conditional gradient methods under interpolation-like conditionsTraining Neural Networks as Learning Data-adaptive Kernels: Provable Representation and Approximation BenefitsHARFE: hard-ridge random feature expansionOn the Inconsistency of Kernel Ridgeless Regression in Fixed DimensionsA Universal Trade-off Between the Model Size, Test Loss, and Training Loss of Linear PredictorsSVRG meets AdaGrad: painless variance reductionUnnamed ItemBenign Overfitting and Noisy FeaturesTractability from overparametrization: the example of the negative perceptronUnnamed ItemUnnamed ItemLinearized two-layers neural networks in high dimensionDiversity Sampling is an Implicit Regularization for Kernel MethodsGeneralization Error of Minimum Weighted Norm and Kernel InterpolationUnnamed ItemOn the robustness of minimum norm interpolators and regularized empirical risk minimizersA Multi-resolution Theory for Approximating Infinite-p-Zero-n: Transitional Inference, Individualized Predictions, and a World Without Bias-Variance TradeoffUnnamed ItemUnnamed ItemMultilevel Fine-Tuning: Closing Generalization Gaps in Approximation of Solution Maps under a Limited Budget for Training DataA Unifying Tutorial on Approximate Message PassingThe interpolation phase transition in neural networks: memorization and generalization under lazy trainingA sieve stochastic gradient descent estimator for online nonparametric regression in Sobolev ellipsoidsA random matrix analysis of random Fourier features: beyond the Gaussian kernel, a precise phase transition, and the corresponding double descent*For interpolating kernel machines, minimizing the norm of the ERM solution maximizes stability


Uses Software


Cites Work


This page was built for publication: Just interpolate: kernel ``ridgeless regression can generalize