3D scene understanding in two glances

This project aims to understand how adults and infants segment visual scenes by developing new display technologies and using machine learning to simulate and analyze 3D vision inputs.

Subsidie
€ 2.126.444
2023

Projectdetails

Introduction

The human mind understands visual scenes. We can usually tell what objects are present in a scene, we can imagine what the hidden parts of objects look like, and we can imagine what it would look like if we or an object moved.

Segmentation in Visual Understanding

The first step of visual scene understanding is segmentation, in which our brain tries to infer which parts of the scene belong to which objects. Adults can do this in photographs – but photographs are not how we learned to see as infants. We learned to see by moving around in a 3D world.

Information Processing

The way that scenes project into our eyes, how light is affected by the optics of our eyes, how our photoreceptors sample the light, and how we move our eyes all provide rich information about our environment. However, we do not know how adults combine all this information to perceive segmented scenes, and we do not know how infants learn this combination.

Challenges in Research

Two reasons for this are:

  1. Standard visual display devices cannot precisely mimic these factors.
  2. It is unethical to manipulate these factors in human infants.

Project Goals

The goals of this project are to:

  • Understand how adults use the rich information present in active 3D vision to perform segmentation.
  • Understand how this is learned.

Methodology

We will develop a new display device and experimental methods to study how adults segment scenes when realistic visual information is available. Additionally, we will develop groundbreaking new technologies using advanced computer graphics and machine learning to simulate the inputs to the visual system from early development to adulthood.

Experiments and Comparisons

We will then conduct in silico experiments in artificial neural networks to understand segmentation learning by systematically restricting or manipulating different factors. We will compare the learned behaviors of different artificial networks to adults performing segmentation during active exploration of 3D scenes.

Conclusion

We will use similarities and differences to better understand a fundamental puzzle of perception: how the mind makes sense of scenes.

Financiële details & Tijdlijn

Financiële details

Subsidiebedrag€ 2.126.444
Totale projectbegroting€ 2.126.444

Tijdlijn

Startdatum1-11-2023
Einddatum31-10-2028
Subsidiejaar2023

Partners & Locaties

Projectpartners

  • TECHNISCHE UNIVERSITAT DARMSTADTpenvoerder

Land(en)

Germany

Vergelijkbare projecten binnen European Research Council

ERC Starting...

Spatial 3D Semantic Understanding for Perception in the Wild

The project aims to develop new algorithms for robust 3D visual perception and semantic understanding from 2D images, enhancing machine perception and immersive technologies.

€ 1.500.000
ERC Starting...

It's about time: Towards a dynamic account of natural vision.

TIME aims to revolutionize vision research by integrating semantic understanding and active information sampling through advanced brain imaging and bio-inspired deep learning, enhancing insights into visual cognition.

€ 1.499.455
ERC Starting...

Using deep neural networks to understand functional specialization in the human visual cortex

This project aims to uncover the origins of functional specialization in the brain's visual pathway by integrating computational modeling, naturalistic behavior sampling, and neuroimaging.

€ 1.494.750
ERC Starting...

Detailed Cortical Mechanisms of Top-Down Visual Processing

This project aims to explore the neural mechanisms of generative vision in macaque monkeys using advanced imaging and behavioral tasks, linking findings to artificial intelligence and human perception.

€ 1.500.000
ERC Consolid...

A perturbative approach to model retinal processing of natural scenes

This project aims to develop realistic deep network models to understand retinal processing of natural scenes by mapping model components to retinal cell types and probing selectivity to stimuli perturbations.

€ 1.998.280