Event-based Object Segmentation
7 papers with code • 4 benchmarks • 4 datasets
Most implemented papers
Segment Anything
We introduce the Segment Anything (SA) project: a new task, model, and dataset for image segmentation.
High Speed and High Dynamic Range Video with an Event Camera
In this work we propose to learn to reconstruct intensity images from event streams directly from data instead of relying on any hand-crafted priors.
Event-Based Video Reconstruction Using Transformer
Event cameras, which output events by detecting spatio-temporal brightness changes, bring a novel paradigm to image sensors with high dynamic range and low latency.
Dual Transfer Learning for Event-based End-task Prediction via Pluggable Event to Image Translation
Event cameras are novel sensors that perceive the per-pixel intensity changes and output asynchronous event streams with high dynamic range and less motion blur.
EvDistill: Asynchronous Events to End-task Learning via Bidirectional Reconstruction-guided Cross-modal Knowledge Distillation
To enable KD across the unpaired modalities, we first propose a bidirectional modality reconstruction (BMR) module to bridge both modalities and simultaneously exploit them to distill knowledge via the crafted pairs, causing no extra computation in the inference.
ESS: Learning Event-based Semantic Segmentation from Still Images
Nonetheless, semantic segmentation with event cameras is still in its infancy which is chiefly due to the lack of high-quality, labeled datasets.
Segment Any Events via Weighted Adaptation of Pivotal Tokens
One pivotal issue at the heart of this endeavor is the precise alignment and calibration of embeddings derived from event-centric data such that they harmoniously coincide with those originating from RGB imagery.