AVECL-UMONS database for audio-visual event classification and localization

2 Oct 2020  ·  Mathilde Brousmiche, Stéphane Dupont, Jean Rouat ·

We introduce the AVECL-UMons dataset for audio-visual event classification and localization in the context of office environments. The audio-visual dataset is composed of 11 event classes recorded at several realistic positions in two different rooms. Two types of sequences are recorded according to the number of events in the sequence. The dataset comprises 2662 unilabel sequences and 2724 multilabel sequences corresponding to a total of 5.24 hours. The dataset is publicly accessible online : https://zenodo.org/record/3965492#.X09wsobgrCI.

PDF Abstract
No code implementations yet. Submit your code now

Datasets


Introduced in the Paper:

AVECL-UMons

Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods


No methods listed for this paper. Add relevant methods here