LSSED: a large-scale dataset and benchmark for speech emotion recognition

30 Jan 2021  ·  Weiquan Fan, Xiangmin Xu, Xiaofen Xing, Weidong Chen, DongYan Huang ·

Speech emotion recognition is a vital contributor to the next generation of human-computer interaction (HCI). However, current existing small-scale databases have limited the development of related research. In this paper, we present LSSED, a challenging large-scale english speech emotion dataset, which has data collected from 820 subjects to simulate real-world distribution. In addition, we release some pre-trained models based on LSSED, which can not only promote the development of speech emotion recognition, but can also be transferred to related downstream tasks such as mental health analysis where data is extremely difficult to collect. Finally, our experiments show the necessity of large-scale datasets and the effectiveness of pre-trained models. The dateset will be released on https://github.com/tobefans/LSSED.

PDF Abstract

Datasets


Introduced in the Paper:

LSSED

Used in the Paper:

IEMOCAP MELD

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Speech Emotion Recognition LSSED PyResNet Unweighted Accuracy (UA) 0.429 # 1

Methods


No methods listed for this paper. Add relevant methods here