SoccerNet-v2: A Dataset and Benchmarks for Holistic Understanding of Broadcast Soccer Videos

Understanding broadcast videos is a challenging task in computer vision, as it requires generic reasoning capabilities to appreciate the content offered by the video editing. In this work, we propose SoccerNet-v2, a novel large-scale corpus of manual annotations for the SoccerNet video dataset, along with open challenges to encourage more research in soccer understanding and broadcast production. Specifically, we release around 300k annotations within SoccerNet's 500 untrimmed broadcast soccer videos. We extend current tasks in the realm of soccer to include action spotting, camera shot segmentation with boundary detection, and we define a novel replay grounding task. For each task, we provide and discuss benchmark results, reproducible with our open-source adapted implementations of the most relevant works in the field. SoccerNet-v2 is presented to the broader research community to help push computer vision closer to automatic solutions for more general video understanding and production purposes.

PDF Abstract

Datasets


Introduced in the Paper:

SoccerNet-v2

Used in the Paper:

ActivityNet THUMOS14 SoccerNet SoccerDB
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Camera shot segmentation SoccerNet-v2 CALF (Cioppa et al.) mIoU 47.3 # 1
Replay Grounding SoccerNet-v2 CALF (Cioppa et al.) Average-AP 41.8 # 1
Replay Grounding SoccerNet-v2 NetVLAD (Giancola et al.) Average-AP 24.3 # 2
Camera shot boundary detection SoccerNet-v2 Histogram (Scikit-Video) mAP 78.5 # 1
Camera shot boundary detection SoccerNet-v2 Content (PySceneDetect) mAP 62.2 # 3
Camera shot boundary detection SoccerNet-v2 Intensity (Scikit-Video) mAP 64.0 # 2
Camera shot boundary detection SoccerNet-v2 CALF (Cioppa et al.) mAP 59.6 # 4
Camera shot segmentation SoccerNet-v2 Baseline mIoU 35.8 # 2
Action Spotting SoccerNet-v2 NetVLAD (Giancola et al.) Average-mAP 31.4 # 10
Action Spotting SoccerNet-v2 AudioVid (Vanderplaetse et al.) Average-mAP 39.9 # 9

Methods


No methods listed for this paper. Add relevant methods here