Neural OmniVideo: Fusing World Knowledge into Smart Video-Specific Models
Develop Neural OmniVideo Models to enhance video analysis and synthesis by integrating deep learning frameworks with external knowledge for improved representation and understanding of dynamic content.
Projectdetails
Introduction
The field of computer vision has made unprecedented progress in applying Deep Learning (DL) to images. Nevertheless, expanding this progress to videos is dramatically lagging behind, due to two key challenges:
- Video data is highly complex and diverse, requiring an order of magnitude more training data than images.
- Raw video data is extremely high dimensional.
These challenges make the processing of entire video pixel-volumes at scale prohibitively expensive and ineffective. Thus, applying DL at scale to video is restricted to short clips or aggressively sub-sampled videos.
Video-Specific Models
On the other side of the spectrum, video-specific models—a single or a few neural networks trained on a single video—exhibit several key properties:
- Facilitate effective video representations (e.g., layers) that make video analysis and editing significantly more tractable.
- Enable long-range temporal analysis by encoding the video through the network.
- Are not restricted to the distribution of training data.
Nevertheless, the capabilities, applicability, and robustness of such models are hampered by having access to only low-level information in the video.
Proposed Solution
We propose to combine the power of these two approaches by the new concept of Neural OmniVideo Models: DL-based frameworks that effectively represent the dynamics of a given video, coupled with the vast knowledge learned by an ensemble of external models.
Objectives
We are aimed at pioneering novel methodologies for developing such models for video analysis and synthesis tasks. Our approach will have several important outcomes:
- Give rise to fundamentally novel effective video representations.
- Go beyond state-of-the-art in classical video analysis tasks that involve long-range temporal analysis.
- Enhance the perception of our dynamic world through new synthesis capabilities.
- Gain profound understanding of the internal representation learned by state-of-the-art large-scale models, and unveil new priors about our dynamic world.
Financiële details & Tijdlijn
Financiële details
Subsidiebedrag | € 1.500.000 |
Totale projectbegroting | € 1.500.000 |
Tijdlijn
Startdatum | 1-2-2024 |
Einddatum | 31-1-2029 |
Subsidiejaar | 2024 |
Partners & Locaties
Projectpartners
- WEIZMANN INSTITUTE OF SCIENCEpenvoerder
Land(en)
Vergelijkbare projecten binnen European Research Council
Project | Regeling | Bedrag | Jaar | Actie |
---|---|---|---|---|
Omni-Supervised Learning for Dynamic Scene UnderstandingThis project aims to enhance dynamic scene understanding in autonomous vehicles by developing innovative machine learning models and methods for open-world object recognition from unlabeled video data. | ERC Starting... | € 1.500.000 | 2023 | Details |
Federated foundational models for embodied perceptionThe FRONTIER project aims to develop foundational models for embodied perception by integrating neural networks with physical simulations, enhancing learning efficiency and collaboration across intelligent systems. | ERC Advanced... | € 2.499.825 | 2024 | Details |
Exploration of Unknown Environments for Digital TwinsThe 'explorer' project aims to automate video data capture and labeling in open worlds to facilitate the creation of semantically rich Digital Twins for complex environments using AI-driven methods. | ERC Advanced... | € 2.476.718 | 2023 | Details |
Learning to synthesize interactive 3D modelsThis project aims to automate the generation of interactive 3D models using deep learning to enhance virtual environments and applications in animation, robotics, and digital entertainment. | ERC Consolid... | € 2.000.000 | 2024 | Details |
Dynamics-Aware Theory of Deep LearningThis project aims to create a robust theoretical framework for deep learning, enhancing understanding and practical tools to improve model performance and reduce complexity in various applications. | ERC Starting... | € 1.498.410 | 2022 | Details |
Omni-Supervised Learning for Dynamic Scene Understanding
This project aims to enhance dynamic scene understanding in autonomous vehicles by developing innovative machine learning models and methods for open-world object recognition from unlabeled video data.
Federated foundational models for embodied perception
The FRONTIER project aims to develop foundational models for embodied perception by integrating neural networks with physical simulations, enhancing learning efficiency and collaboration across intelligent systems.
Exploration of Unknown Environments for Digital Twins
The 'explorer' project aims to automate video data capture and labeling in open worlds to facilitate the creation of semantically rich Digital Twins for complex environments using AI-driven methods.
Learning to synthesize interactive 3D models
This project aims to automate the generation of interactive 3D models using deep learning to enhance virtual environments and applications in animation, robotics, and digital entertainment.
Dynamics-Aware Theory of Deep Learning
This project aims to create a robust theoretical framework for deep learning, enhancing understanding and practical tools to improve model performance and reduce complexity in various applications.
Vergelijkbare projecten uit andere regelingen
Project | Regeling | Bedrag | Jaar | Actie |
---|---|---|---|---|
VID-AI: VIDeo feAture detectIon and retrievalSensity en BrainCreators ontwikkelen een geavanceerd platform voor het detecteren van visuele bedreigingen, zoals deepfakes, door video-embedding en retrieval technologieën te integreren. | Mkb-innovati... | € 248.220 | 2020 | Details |
Een standaard voor productiewaardige Deep Learning systemenHet project richt zich op het verbeteren van audio- en video-analyse systemen door samenwerking tussen Media Distillery, NovoLanguage en een partner, met als doel hogere kwaliteit en snellere ontwikkeling via gedeelde technologieën. | Mkb-innovati... | € 104.061 | 2016 | Details |
VID-AI: VIDeo feAture detectIon and retrieval
Sensity en BrainCreators ontwikkelen een geavanceerd platform voor het detecteren van visuele bedreigingen, zoals deepfakes, door video-embedding en retrieval technologieën te integreren.
Een standaard voor productiewaardige Deep Learning systemen
Het project richt zich op het verbeteren van audio- en video-analyse systemen door samenwerking tussen Media Distillery, NovoLanguage en een partner, met als doel hogere kwaliteit en snellere ontwikkeling via gedeelde technologieën.