We introduce a new dataset, Watch and Learn Time-lapse (WALT), consisting of multiple (4K and 1080p) cameras capturing urban environments over a year.
7 PAPERS • 1 BENCHMARK
MOViD-A is a video-based synthesized dataset. We create it from MOVi dataset for amodal segmentation. The virtual camera is set to go around the scene, capturing about 24 consecutive frames. We randomly place 10 ∼ 20 static objects that heavily occlude each other in the scene. Finally, we collect 630 and 208 videos for training and testing.
1 PAPER • NO BENCHMARKS YET