Emotion Recognition

467 papers with code • 7 benchmarks • 45 datasets

Emotion Recognition is an important area of research to enable effective human-computer interaction. Human emotions can be detected using speech signal, facial expressions, body language, and electroencephalography (EEG). Source: Using Deep Autoencoders for Facial Expression Recognition

Latest papers with no code

Self-supervised Gait-based Emotion Representation Learning from Selective Strongly Augmented Skeleton Sequences

no code yet • 8 May 2024

In this paper, we propose a contrastive learning framework utilizing selective strong augmentation (SSA) for self-supervised gait-based emotion representation, which aims to derive effective representations from limited labeled gait data.

Empathy Through Multimodality in Conversational Interfaces

no code yet • 8 May 2024

Agents represent one of the most emerging applications of Large Language Models (LLMs) and Generative AI, with their effectiveness hinging on multimodal capabilities to navigate complex user environments.

Adapting WavLM for Speech Emotion Recognition

no code yet • 7 May 2024

Recently, the usage of speech self-supervised models (SSL) for downstream tasks has been drawing a lot of attention.

ESIHGNN: Event-State Interactions Infused Heterogeneous Graph Neural Network for Conversational Emotion Recognition

no code yet • 7 May 2024

Toward this end, we propose a novel graph-based approach, namely Event-State Interactions infused Heterogeneous Graph Neural Network (ESIHGNN), which incorporates the speaker's emotional state and constructs a heterogeneous event-state interaction graph to model the conversation.

GMP-ATL: Gender-augmented Multi-scale Pseudo-label Enhanced Adaptive Transfer Learning for Speech Emotion Recognition via HuBERT

no code yet • 3 May 2024

The continuous evolution of pre-trained speech models has greatly advanced Speech Emotion Recognition (SER).

Toward end-to-end interpretable convolutional neural networks for waveform signals

no code yet • 3 May 2024

This paper introduces a novel convolutional neural networks (CNN) framework tailored for end-to-end audio deep learning models, presenting advancements in efficiency and explainability.

Converting Anyone's Voice: End-to-End Expressive Voice Conversion with a Conditional Diffusion Model

no code yet • 2 May 2024

A major challenge of expressive VC lies in emotion prosody modeling.

EALD-MLLM: Emotion Analysis in Long-sequential and De-identity videos with Multi-modal Large Language Model

no code yet • 1 May 2024

In contrast, long sequential videos can reveal authentic emotions; 2) Previous studies commonly utilize various signals such as facial, speech, and even sensitive biological signals (e. g., electrocardiogram).

Usefulness of Emotional Prosody in Neural Machine Translation

no code yet • 27 Apr 2024

In this work, we propose to improve translation quality by adding another external source of information: the automatically recognized emotion in the voice.

Revisiting Multimodal Emotion Recognition in Conversation from the Perspective of Graph Spectrum

no code yet • 27 Apr 2024

Since consistency and complementarity information correspond to low-frequency and high-frequency information, respectively, this paper revisits the problem of multimodal emotion recognition in conversation from the perspective of the graph spectrum.