Contrast and Generation Make BART a Good Dialogue Emotion Recognizer

21 Dec 2021  ·  ShiMin Li, Hang Yan, Xipeng Qiu ·

In dialogue systems, utterances with similar semantics may have distinctive emotions under different contexts. Therefore, modeling long-range contextual emotional relationships with speaker dependency plays a crucial part in dialogue emotion recognition. Meanwhile, distinguishing the different emotion categories is non-trivial since they usually have semantically similar sentiments. To this end, we adopt supervised contrastive learning to make different emotions mutually exclusive to identify similar emotions better. Meanwhile, we utilize an auxiliary response generation task to enhance the model's ability of handling context information, thereby forcing the model to recognize emotions with similar semantics in diverse contexts. To achieve these objectives, we use the pre-trained encoder-decoder model BART as our backbone model since it is very suitable for both understanding and generation tasks. The experiments on four datasets demonstrate that our proposed model obtains significantly more favorable results than the state-of-the-art model in dialogue emotion recognition. The ablation study further demonstrates the effectiveness of supervised contrastive loss and generative loss.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Emotion Recognition in Conversation DailyDialog CoG-BART Micro-F1 54.71 # 16
Weighted F1 54.71 # 1
Emotion Recognition in Conversation EmoryNLP CoG-BART Weighted-F1 39.04 # 11
Micro-F1 42.58 # 3
Emotion Recognition in Conversation MELD CoG-BART Weighted-F1 64.81 # 31

Methods