Breaking the Curse of Dimensionality with Convex Neural Networks

From MaRDI portal
Publication:5361282

zbMath1433.68390arXiv1412.8690MaRDI QIDQ5361282

Francis Bach

Publication date: 27 September 2017

Full work available at URL: https://arxiv.org/abs/1412.8690




Related Items

Mehler’s Formula, Branching Process, and Compositional Kernels of Deep Neural NetworksA Cross-Validated Ensemble Approach to Robust Hypothesis Testing of Continuous Nonlinear Interactions: Application to Nutrition-Environment StudiesA Regularity Theory for Static Schrödinger Equations on \(\boldsymbol{\mathbb{R}}\)d in Spectral Barron SpacesDeep learning: a statistical viewpointNeural network approximationSpurious Valleys in Two-layer Neural Network Optimization LandscapesA Proof that Artificial Neural Networks Overcome the Curse of Dimensionality in the Numerical Approximation of Black–Scholes Partial Differential EquationsMachine learning from a continuous viewpoint. IChallenges in optimization with complex PDE-systems. Abstracts from the workshop held February 14--20, 2021 (hybrid meeting)Thermodynamically consistent physics-informed neural networks for hyperbolic systemsDegrees of freedom for off-the-grid sparse estimationTwo-Layer Neural Networks with Values in a Banach SpaceParticle dual averaging: optimization of mean field neural network with global convergence rate analysis*Locality defeats the curse of dimensionality in convolutional teacher–student scenarios*Relative stability toward diffeomorphisms indicates performance in deep nets*Finite-Sample Two-Group Composite Hypothesis Testing via Machine LearningA proof of convergence for gradient descent in the training of artificial neural networks for constant target functionsA precise high-dimensional asymptotic theory for boosting and minimum-\(\ell_1\)-norm interpolated classifiersSparse optimization on measures with over-parameterized gradient descentDeep learning for constrained utility maximisationNonlinear Variable Selection via Deep Neural NetworksApproximation properties of deep ReLU CNNsUniform approximation rates and metric entropy of shallow neural networksWhat Kinds of Functions Do Deep Neural Networks Learn? Insights from Variational Spline TheoryBenign overfitting in linear regressionUnnamed ItemDeep ReLU Networks Overcome the Curse of Dimensionality for Generalized Bandlimited FunctionsNonconvex regularization for sparse neural networksFull error analysis for the training of deep neural networksTwo Steps at a Time---Taking GAN Training in Stride with Tseng's MethodLearning the mapping \(\mathbf{x}\mapsto \sum\limits_{i=1}^d x_i^2\): the cost of finding the needle in a haystackBayesian Imaging Using Plug & Play Priors: When Langevin Meets TweedieTraining Neural Networks as Learning Data-adaptive Kernels: Provable Representation and Approximation BenefitsConditional regression for single-index modelsPiecewise linear functions representable with infinite width shallow ReLU neural networksRates of approximation by ReLU shallow neural networksProbabilistic partition of unity networks for high‐dimensional regression problemsDEEP EQUILIBRIUM NETSDeep Learning for Marginal Bayesian Posterior Inference with Recurrent Neural NetworksA priori generalization error analysis of two-layer neural networks for solving high dimensional Schrödinger eigenvalue problemsQuantum Monte Carlo for economics: stress testing and macroeconomic deep learningFunction approximation by deep networksOverall error analysis for the training of deep neural networks via stochastic gradient descent with random initialisationConvergence rates of gradient methods for convex optimization in the space of measuresA mathematical perspective of machine learningUnnamed ItemUnnamed ItemThe geometry of off-the-grid compressed sensingSelf-Supervised Deep Learning for Image Reconstruction: A Langevin Monte Carlo ApproachEstimating Optimal Infinite Horizon Dynamic Treatment Regimes via pT-LearningOn maximum a posteriori estimation with Plug \& Play priors and stochastic gradient descentEfficient Global Optimization of Two-Layer ReLU Networks: Quadratic-Time Algorithms and Adversarial TrainingGradient descent on infinitely wide neural networks: global convergence and generalizationGreedy training algorithms for neural networks and applications to PDEsNonparametric regression using deep neural networks with ReLU activation functionExSpliNet: An interpretable and expressive spline-based neural networkKolmogorov width decay and poor approximators in machine learning: shallow neural networks, random feature models and neural tangent kernelsLandscape and training regimes in deep learningDeep Neural Networks Algorithms for Stochastic Control Problems on Finite Horizon: Convergence AnalysisUnnamed ItemMachine Learning and Computational MathematicsAnalysis of a two-layer neural network via displacement convexityEnsemble feature selection using election methods and ranker clusteringUnnamed ItemTopological properties of the set of functions generated by neural networks of fixed sizeLinearized two-layers neural networks in high dimensionHigh-dimensional index volatility models via Stein's identityError bounds for deep ReLU networks using the Kolmogorov-Arnold superposition theoremFast generalization error bound of deep learning without scale invariance of activation functionsOptimally weighted loss functions for solving PDEs with neural networksFitting small piece-wise linear neural network models to interpolate data setsSupervised learning from noisy observations: combining machine-learning techniques with data assimilationMultikernel Regression with Sparsity ConstraintThe Gap between Theory and Practice in Function Approximation with Deep Neural NetworksUnnamed ItemUnnamed ItemAsymptotic learning curves of kernel methods: empirical data versus teacher–student paradigmUnnamed ItemUnnamed ItemCAS4DL: Christoffel adaptive sampling for function approximation via deep learningUnderstanding neural networks with reproducing kernel Banach spacesThe interpolation phase transition in neural networks: memorization and generalization under lazy trainingOn the Effectiveness of Richardson Extrapolation in Data ScienceWhen do neural networks outperform kernel methods?*Approximation Error Analysis of Some Deep Backward Schemes for Nonlinear PDEsRepresentation formulas and pointwise properties for Barron functionsDeep ReLU neural networks overcome the curse of dimensionality for partial integrodifferential equationsApproximating functions with multi-features by deep convolutional neural networksThe Barron space and the flow-induced function spaces for neural network modelsRobust and resource-efficient identification of two hidden layer neural networksHigh-order approximation rates for shallow neural networks with cosine and \(\mathrm{ReLU}^k\) activation functions