Action Unit Detection

14 papers with code • 1 benchmarks • 3 datasets

Action unit detection is the task of detecting action units from a video - for example, types of facial action units (lip tightening, cheek raising) from a video of a face.

( Image credit: AU R-CNN )

Libraries

Use these libraries to find Action Unit Detection models and implementations

Most implemented papers

J$\hat{\text{A}}$A-Net: Joint Facial Action Unit Detection and Face Alignment via Adaptive Attention

ZhiwenShao/PyTorch-JAANet 18 Mar 2020

Moreover, to extract precise local features, we propose an adaptive attention learning module to refine the attention map of each AU adaptively.

EmotiEffNet Facial Features in Uni-task Emotion Recognition in Video at ABAW-5 competition

HSE-asavchenko/face-emotion-recognition 16 Mar 2023

In this article, the results of our team for the fifth Affective Behavior Analysis in-the-wild (ABAW) competition are presented.

FG-Net: Facial Action Unit Detection with Generalizable Pyramidal Features

ihp-lab/fg-net 23 Aug 2023

The proposed FG-Net achieves a strong generalization ability for heatmap-based AU detection thanks to the generalizable and semantic-rich features extracted from the pre-trained generative model.

Learning Contrastive Feature Representations for Facial Action Unit Detection

ziqiao-shang/aunce 9 Feb 2024

To address the challenge posed by noisy AU labels, we augment the supervised signal through the introduction of a self-supervised signal.