Optimizing for Generalization in Machine Learning

This project aims to unravel the mystery of generalization in machine learning by developing novel optimization algorithms to enhance the reliability and applicability of ML in critical domains.

Subsidie
€ 1.494.375
2023

Projectdetails

Introduction

Recent advances in the field of machine learning (ML) are revolutionizing an ever-growing variety of domains, ranging from statistical learning algorithms in computer vision and natural language processing all the way to reinforcement learning algorithms in autonomous driving and conversational AI.

The Generalization Mystery

However, many of these breakthroughs demonstrate phenomena that lack explanations, and sometimes even contradict conventional wisdom. Perhaps the greatest mystery of modern ML—and arguably, one of the greatest mysteries of all of modern computer science—is the question of generalization: why do these immensely complex prediction rules successfully apply to future unseen instances?

Importance of Understanding Generalization

Apart from the pure scientific curiosity it stimulates, I believe that this lack of understanding poses a significant obstacle to widening the applicability of ML to critical applications, like in healthcare or autonomous driving, where the cost of error is disastrous.

Project Goals

The broad goal of this project is to tackle the generalization mystery in the context of both statistical learning and reinforcement learning, focusing on optimization algorithms being the de facto contemporary standard in training learning models.

Methodology

Our methodology points out inherent shortcomings of widely accepted viewpoints with regard to the generalization of optimization-based learning algorithms. It takes a crucially different approach that targets the optimization algorithm itself.

Key Steps

  1. Building bottom-up from fundamental and tractable optimization models.
  2. Identifying intrinsic properties.
  3. Developing algorithmic methodologies that enable optimization to effectively generalize in modern statistical- and reinforcement-learning scenarios.

Expected Outcomes

A successful outcome would not only lead to a timely and crucial shift in the way the research community approaches the generalization of contemporary optimization-based ML, but it may also significantly transform the way we develop practical, efficient, and reliable learning systems.

Financiële details & Tijdlijn

Financiële details

Subsidiebedrag€ 1.494.375
Totale projectbegroting€ 1.494.375

Tijdlijn

Startdatum1-10-2023
Einddatum30-9-2028
Subsidiejaar2023

Partners & Locaties

Projectpartners

  • TEL AVIV UNIVERSITYpenvoerder

Land(en)

Israel

Vergelijkbare projecten binnen European Research Council

ERC Starting...

Foundations of Generalization

This project aims to explore generalization in overparameterized learning models through stochastic convex optimization and synthetic data generation, enhancing understanding of modern algorithms.

€ 1.419.375
ERC Starting...

Modern Challenges in Learning Theory

This project aims to develop a new theory of generalization in machine learning that better models real-world tasks and addresses data efficiency and privacy challenges.

€ 1.433.750
ERC Consolid...

Theoretical Understanding of Classic Learning Algorithms

The TUCLA project aims to enhance classic machine learning algorithms, particularly Bagging and Boosting, to achieve faster, data-efficient learning and improve their theoretical foundations.

€ 1.999.288
ERC Starting...

Uniting Statistical Testing and Machine Learning for Safe Predictions

The project aims to enhance the interpretability and reliability of machine learning predictions by integrating statistical methods to establish robust error bounds and ensure safe deployment in real-world applications.

€ 1.500.000
ERC Starting...

Algorithmic Bias Control in Deep learning

The project aims to develop a theory of algorithmic bias in deep learning to improve training efficiency and generalization performance for real-world applications.

€ 1.500.000