Multilogue-Net: A Context Aware RNN for Multi-modal Emotion Detection and Sentiment Analysis in Conversation

arXiv preprint 2020  ·  Aman Shenoy, Ashish Sardana ·

Sentiment Analysis and Emotion Detection in conversation is key in several real-world applications, with an increase in modalities available aiding a better understanding of the underlying emotions. Multi-modal Emotion Detection and Sentiment Analysis can be particularly useful, as applications will be able to use specific subsets of available modalities, as per the available data. Current systems dealing with Multi-modal functionality fail to leverage and capture - the context of the conversation through all modalities, the dependency between the listener(s) and speaker emotional states, and the relevance and relationship between the available modalities. In this paper, we propose an end to end RNN architecture that attempts to take into account all the mentioned drawbacks. Our proposed model, at the time of writing, out-performs the state of the art on a benchmark dataset on a variety of accuracy and regression metrics.

PDF Abstract arXiv preprint 2020 PDF arXiv preprint 2020 Abstract

Datasets


Results from the Paper


Ranked #8 on Multimodal Sentiment Analysis on CMU-MOSEI (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Benchmark
Multimodal Sentiment Analysis CMU-MOSEI Multilogue-Net Accuracy 82.10 # 8
MAE 0.59 # 4

Methods


No methods listed for this paper. Add relevant methods here