Flexible Statistical Inference

Develop a flexible statistical theory allowing post-hoc data collection and decision-making with error control, utilizing e-values for improved inference in small samples.

Subsidie
€ 2.499.461
2024

Projectdetails

Introduction

Most statistical methods require that all aspects of data collection and inference are determined in advance, independently of the data. These include when to stop collecting data, what decisions can be made (e.g., accept/reject hypothesis, classify new point), and how to measure their quality (e.g., loss function/significance level). This is wildly at odds with the flexibility required in practice! It makes it impossible to achieve error control in meta-analyses and contributes to the replication crisis in the applied sciences.

Project Overview

I will develop a novel statistical theory in which all data-collection and decision aspects may be unknown in advance, possibly imposed post-hoc, depending on the data itself in unknowable ways. Yet this new theory will provide small-sample frequentist error control, risk bounds, and confidence sets.

Theoretical Foundations

I base myself on far-reaching extensions of e-values/processes. These generalize likelihood ratios and replace p-values, capturing 'evidence' in a much cleaner fashion. As lead author of the first paper (2019) that gave e-values a name and demonstrated their enormous potential, I kicked off and then played an essential role in the extremely rapid development of anytime-valid inference, the one aspect of flexibility that is by now well-studied.

Research Goals

Still, efficient e-value design principles for many standard problems (e.g., GLMs and other settings with covariates) are still lacking, and I will provide them. I will also develop theory for full decision-task flexibility, about which currently almost nothing is known. A major innovation is the e-posterior, which behaves differently from the Bayesian one: if priors are chosen badly, e-posterior based confidence intervals get wide rather than wrong.

Implications of the New Theory

Both the existing Wald-Neyman-Pearson and Bayesian statistical theories will arise as special, extreme cases of the new theory, based on perfect (hence unrealistic) knowledge of the data-collection/decision problem or the underlying distribution(s), respectively.

Financiële details & Tijdlijn

Financiële details

Subsidiebedrag€ 2.499.461
Totale projectbegroting€ 2.499.461

Tijdlijn

Startdatum1-11-2024
Einddatum31-10-2029
Subsidiejaar2024

Partners & Locaties

Projectpartners

  • STICHTING NEDERLANDSE WETENSCHAPPELIJK ONDERZOEK INSTITUTENpenvoerder

Land(en)

Netherlands

Vergelijkbare projecten binnen European Research Council

ERC Advanced...

Assumption-Lean (Causal) Modelling and Estimation: A Paradigm Shift from Traditional Statistical Modelling

Develop a flexible 'assumption-lean modelling' framework for causal inference that minimizes bias and enhances interpretability in statistical analyses using debiased learning techniques.

€ 2.445.063
ERC Starting...

The missing mathematical story of Bayesian uncertainty quantification for big data

This project aims to enhance scalable Bayesian methods through theoretical insights, improving their accuracy and acceptance in real-world applications like medicine and cosmology.

€ 1.492.750
ERC Starting...

Information Economics With Fundamental Uncertainty: Robustness, Commitment, and Strategic Incentives

This project explores how fundamental uncertainty affects learning and strategic information sharing, using decision-theoretic frameworks to enhance understanding of active experimentation and information dynamics.

€ 1.402.678
ERC Starting...

Statistical theory and methodology for the combination of heterogeneous and distributed data

Develop new statistical methodologies to address data heterogeneity and measurement errors across diverse datasets, enhancing evidence-based advancements in science and policy.

€ 1.499.689
ERC Starting...

Uniting Statistical Testing and Machine Learning for Safe Predictions

The project aims to enhance the interpretability and reliability of machine learning predictions by integrating statistical methods to establish robust error bounds and ensure safe deployment in real-world applications.

€ 1.500.000