Landscape and training regimes in deep learning
From MaRDI portal
Publication:2231925
DOI10.1016/j.physrep.2021.04.001OpenAlexW3153303803MaRDI QIDQ2231925
Mario Geiger, Matthieu Wyart, Leonardo Petrini
Publication date: 30 September 2021
Published in: Physics Reports (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.physrep.2021.04.001
neural networkscurse of dimensionalityjammingfeature learningdeep learningneural tangent kernellazy trainingloss landscape
Related Items (1)
Uses Software
Cites Work
- Unnamed Item
- Bayesian learning for neural networks
- Linearized two-layers neural networks in high dimension
- High-dimensional dynamics of generalization error in neural networks
- Mean field analysis of neural networks: a central limit theorem
- The simplest model of jamming
- Exact theory of dense amorphous hard spheres in high dimension. III. The full replica symmetry breaking solution
- A mean field view of the landscape of two-layer neural networks
- Global Minima of Overparameterized Neural Networks
- Reconciling modern machine-learning practice and the classical bias–variance trade-off
- Mean Field Analysis of Neural Networks: A Law of Large Numbers
- Universality of jamming of nonspherical particles
- Breaking the Curse of Dimensionality with Convex Neural Networks
- Comparing dynamics: deep neural networks versus glassy systems
- On the information bottleneck theory of deep learning
- Scaling description of generalization with number of parameters in deep learning
- Disentangling feature and lazy training in deep neural networks
- Asymptotic learning curves of kernel methods: empirical data versus teacher–student paradigm
- Wide neural networks of any depth evolve as linear models under gradient descent *
- Surfing on minima of isostatic landscapes: avalanches and unjamming transition
- A jamming transition from under- to over-parametrization affects generalization in deep learning
This page was built for publication: Landscape and training regimes in deep learning