Multimodal sentiment analysis is the task of performing sentiment analysis with multiple data sources - e.g. a camera feed of someone's face and their recorded speech.
|Trend||Dataset||Best Method||Paper title||Paper||Code||Compare|
Emotion recognition in conversations is crucial for building empathetic machines.
#2 best model for Emotion Recognition in Conversation on IEMOCAP
Humans convey their intentions through the usage of both verbal and nonverbal behaviors during face-to-face communication.
Multimodal sentiment analysis is a developing area of research, which involves the identification of sentiments in videos.
Speech emotion recognition is a challenging task, and extensive reliance has been placed on models that use audio features in building well-performing classifiers.
In this paper, we propose the Gated Multimodal Embedding LSTM with Temporal Attention (GME-LSTM(A)) model that is composed of 2 modules.
AI must understand each modality and the interactions between them that shape human communication.
#3 best model for Multimodal Sentiment Analysis on MOSI
We propose a novel approach to multimodal sentiment analysis using deep neural networks combining visual analysis and natural language processing.
In this paper, we propose a Select-Additive Learning (SAL) procedure that improves the generalizability of trained neural networks for multimodal sentiment analysis.
Our method is based on the key insight that translation from a source to a target modality provides a method of learning joint representations using only the source modality as input.