The Hopkins 155 dataset consists of 156 video sequences of two or three motions. Each video sequence motion corresponds to a low-dimensional subspace. There are 39−550 data vectors drawn from two or three motions for each video sequence.
89 PAPERS • 1 BENCHMARK
ApolloScape is a large dataset consisting of over 140,000 video frames (73 street scene videos) from various locations in China under varying weather conditions. Pixel-wise semantic annotation of the recorded data is provided in 2D, with point-wise semantic annotation in 3D for 28 classes. In addition, the dataset contains lane marking annotations in 2D.
53 PAPERS • 5 BENCHMARKS
Includes accurate pixel-wise motion masks, egomotion and ground truth depth.
11 PAPERS • NO BENCHMARKS YET
A large-scale 4D egocentric dataset with rich annotations, to catalyze the research of category-level human-object interaction. HOI4D consists of 2.4M RGB-D egOCentric video frames over 4000 sequences collected by 4 participants interacting with 800 different object instances from 16 categories over 610 different indoor rooms.
2 PAPERS • NO BENCHMARKS YET
Extension of the official KITTI'15 dataset. independently moving instance segmentation ground truth to cover all moving objects, not just a selection of cars and vans.
1 PAPER • NO BENCHMARKS YET
Please find more details of this dataset at https://alex-xun-xu.github.io/ProjectPage/CVPR_18/index.html
1 PAPER • 1 BENCHMARK