Emotion Recognition

465 papers with code • 7 benchmarks • 45 datasets

Emotion Recognition is an important area of research to enable effective human-computer interaction. Human emotions can be detected using speech signal, facial expressions, body language, and electroencephalography (EEG). Source: Using Deep Autoencoders for Facial Expression Recognition

Active Learning with Task Adaptation Pre-training for Speech Emotion Recognition

clearloveyuan/after 1 May 2024

To address these issues, we propose an active learning (AL)-based fine-tuning framework for SER, called \textsc{After}, that leverages task adaptation pre-training (TAPT) and AL methods to enhance performance and efficiency.

4
01 May 2024

A Systematic Evaluation of Adversarial Attacks against Speech Emotion Recognition Models

limunimi/thesis_adversarial_ml_audio 29 Apr 2024

In summary, this work contributes to the understanding of the robustness of CNN-LSTM models, particularly in SER scenarios, and the impact of AEs.

0
29 Apr 2024

MultiMAE-DER: Multimodal Masked Autoencoder for Dynamic Emotion Recognition

Peihao-Xiang/MultiMAE-DFER 28 Apr 2024

In comparison to state-of-the-art multimodal supervised learning models for dynamic emotion recognition, MultiMAE-DER enhances the weighted average recall (WAR) by 4. 41% on the RAVDESS dataset and by 2. 06% on the CREMAD.

0
28 Apr 2024

MER 2024: Semi-Supervised Learning, Noise Robustness, and Open-Vocabulary Multimodal Emotion Recognition

zeroqiaoba/mertools 26 Apr 2024

In addition to expanding the dataset size, we introduce a new track around open-vocabulary emotion recognition.

104
26 Apr 2024

EmoVIT: Revolutionizing Emotion Insights with Visual Instruction Tuning

aimmemotion/emovit 25 Apr 2024

Visual Instruction Tuning represents a novel learning paradigm involving the fine-tuning of pre-trained language models using task-specific instructions.

3
25 Apr 2024

CAGE: Circumplex Affect Guided Expression Inference

wagner-niklas/cage_expression_inference 23 Apr 2024

Using a small-scaled MaxViT-based model architecture, we evaluate the impact of discrete expression category labels in training with the continuous valence and arousal labels.

3
23 Apr 2024

Cooperative Sentiment Agents for Multimodal Sentiment Analysis

smwanghhh/co-sa 19 Apr 2024

In this paper, we propose a new Multimodal Representation Learning (MRL) method for Multimodal Sentiment Analysis (MSA), which facilitates the adaptive interaction between modalities through Cooperative Sentiment Agents, named Co-SA.

1
19 Apr 2024

MMA-DFER: MultiModal Adaptation of unimodal models for Dynamic Facial Expression Recognition in-the-wild

katerynaCh/MMA-DFER 13 Apr 2024

Within the field of multimodal DFER, recent methods have focused on exploiting advances of self-supervised learning (SSL) for pre-training of strong multimodal encoders.

3
13 Apr 2024

Resolve Domain Conflicts for Generalizable Remote Physiological Measurement

faceonlive/ai-research 11 Apr 2024

Remote photoplethysmography (rPPG) technology has become increasingly popular due to its non-invasive monitoring of various physiological indicators, making it widely applicable in multimedia interaction, healthcare, and emotion analysis.

189
11 Apr 2024

What is Learnt by the LEArnable Front-end (LEAF)? Adapting Per-Channel Energy Normalisation (PCEN) to Noisy Conditions

hanyu-meng/adapting-leaf 10 Apr 2024

There is increasing interest in the use of the LEArnable Front-end (LEAF) in a variety of speech processing systems.

5
10 Apr 2024