VPCD (Video Person-Clustering)

Introduced by Brown et al. in Face, Body, Voice: Video Person-Clustering with Multiple Modalities

VPCD contains multi-modal annotations (face, body and voice) for all primary and secondary characters from a range of diverse TV-shows and movies. It is used for evaluating multi-modal person-clustering. It contains body-tracks for each annotated character, face-tracks when visible, and voice-tracks when speaking, with their associated features.

It consists of more than 30,000 face and body tracks of 300+ characters, from over 23 hours of video.

Papers


Paper Code Results Date Stars

Dataset Loaders


No data loaders found. You can submit your data loader here.

Tasks


Similar Datasets


License


Modalities


Languages