ResEmoteNet: Bridging Accuracy and Loss Reduction in Facial Emotion Recognition

The human face is a silent communicator, expressing emotions and thoughts through its facial expressions. With the advancements in computer vision in recent years, facial emotion recognition technology has made significant strides, enabling machines to decode the intricacies of facial cues. In this work, we propose ResEmoteNet, a novel deep learning architecture for facial emotion recognition designed with the combination of Convolutional, Squeeze-Excitation (SE) and Residual Networks. The inclusion of SE block selectively focuses on the important features of the human face, enhances the feature representation and suppresses the less relevant ones. This helps in reducing the loss and enhancing the overall model performance. We also integrate the SE block with three residual blocks that help in learning more complex representation of the data through deeper layers. We evaluated ResEmoteNet on four open-source databases: FER2013, RAF-DB, AffectNet-7 and ExpW, achieving accuracies of 79.79%, 94.76%, 72.39% and 75.67% respectively. The proposed network outperforms state-of-the-art models across all four databases. The source code for ResEmoteNet is available at https://github.com/ArnabKumarRoy02/ResEmoteNet.

PDF Abstract

Results from the Paper


 Ranked #1 on Facial Expression Recognition (FER) on FER2013 (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Benchmark
Facial Expression Recognition (FER) AffectNet ResEmoteNet Accuracy (7 emotion) 72.93 # 1
Facial Expression Recognition (FER) ExpW ResEmoteNet Accuracy 75.67 # 1
Facial Expression Recognition (FER) FER2013 ResEmoteNet Accuracy 79.79 # 1
Facial Expression Recognition (FER) RAF-DB ResEmoteNet Overall Accuracy 94.76 # 1

Methods


No methods listed for this paper. Add relevant methods here