Understanding Deep Learning
The project aims to establish a solid theoretical foundation for deep learning by investigating optimization, statistical complexity, and representation, enhancing understanding and algorithm development.
Projectdetails
Introduction
While extremely successful, deep learning (DL) still lacks a solid theoretical foundation.
Recent Achievements
In the last 5 years, the PI focused almost entirely on DL theory, yielding a strong publication record with:
- 7 papers at NeurIPS (the leading ML conference), including:
- 2 spotlights (top 3% of submitted papers)
- 1 oral (top 1%)
- 2 papers at ICLR (the leading DL conference)
- 1 paper at COLT (the leading ML theory conference)
These results are amongst the first that break a 20-year hiatus in NN theory, thereby giving some hope for a solid deep learning theory. This includes:
- The first poly-time learnability result for a non-trivial function class by SGD on NN.
- The first such result with near optimal rate.
- New methodology to bound the sample complexity of NN, establishing the first sample complexity bound that is sublinear in the number of parameters, under norm constraints that are valid in practice.
- An explanation for the phenomena of adversarial examples.
Future Plans
We plan to go far beyond these and other results, and to build a coherent theory for DL, addressing all three pillars of learning theory:
Optimization
We plan to investigate the success of SGD in finding a good model, arguably the greatest mystery of modern deep learning. Specifically, our goal is to understand what models are learnable by SGD on neural networks. To this end, we plan to develop a new class of models that can potentially lead to new deep learning algorithms, with a solid theory behind them.
Statistical Complexity
We plan to crack the second great mystery of modern deep learning, which is their ability to generalize with fewer examples than parameters. Our plan is to investigate the sample complexity of classes of neural networks that are defined by bounds on the weights’ magnitude.
Representation
We plan to investigate functions that can be realized by NN. This includes classical questions such as the benefits of depth, as well as more modern aspects such as adversarial examples.
Financiële details & Tijdlijn
Financiële details
Subsidiebedrag | € 1.499.750 |
Totale projectbegroting | € 1.499.750 |
Tijdlijn
Startdatum | 1-9-2022 |
Einddatum | 31-8-2027 |
Subsidiejaar | 2022 |
Partners & Locaties
Projectpartners
- THE HEBREW UNIVERSITY OF JERUSALEMpenvoerder
Land(en)
Vergelijkbare projecten binnen European Research Council
Project | Regeling | Bedrag | Jaar | Actie |
---|---|---|---|---|
Dynamics-Aware Theory of Deep LearningThis project aims to create a robust theoretical framework for deep learning, enhancing understanding and practical tools to improve model performance and reduce complexity in various applications. | ERC Starting... | € 1.498.410 | 2022 | Details |
Algorithmic Bias Control in Deep learningThe project aims to develop a theory of algorithmic bias in deep learning to improve training efficiency and generalization performance for real-world applications. | ERC Starting... | € 1.500.000 | 2022 | Details |
Control for Deep and Federated LearningCoDeFeL aims to enhance machine learning methods through control theory, developing efficient ResNet architectures and federated learning techniques for applications in digital medicine and recommendations. | ERC Advanced... | € 2.499.224 | 2024 | Details |
Modern Challenges in Learning TheoryThis project aims to develop a new theory of generalization in machine learning that better models real-world tasks and addresses data efficiency and privacy challenges. | ERC Starting... | € 1.433.750 | 2022 | Details |
Foundations of GeneralizationThis project aims to explore generalization in overparameterized learning models through stochastic convex optimization and synthetic data generation, enhancing understanding of modern algorithms. | ERC Starting... | € 1.419.375 | 2024 | Details |
Dynamics-Aware Theory of Deep Learning
This project aims to create a robust theoretical framework for deep learning, enhancing understanding and practical tools to improve model performance and reduce complexity in various applications.
Algorithmic Bias Control in Deep learning
The project aims to develop a theory of algorithmic bias in deep learning to improve training efficiency and generalization performance for real-world applications.
Control for Deep and Federated Learning
CoDeFeL aims to enhance machine learning methods through control theory, developing efficient ResNet architectures and federated learning techniques for applications in digital medicine and recommendations.
Modern Challenges in Learning Theory
This project aims to develop a new theory of generalization in machine learning that better models real-world tasks and addresses data efficiency and privacy challenges.
Foundations of Generalization
This project aims to explore generalization in overparameterized learning models through stochastic convex optimization and synthetic data generation, enhancing understanding of modern algorithms.