Learning Skeletal Graph Neural Networks for Hard 3D Pose Estimation

Various deep learning techniques have been proposed to solve the single-view 2D-to-3D pose estimation problem. While the average prediction accuracy has been improved significantly over the years, the performance on hard poses with depth ambiguity, self-occlusion, and complex or rare poses is still far from satisfactory. In this work, we target these hard poses and present a novel skeletal GNN learning solution. To be specific, we propose a hop-aware hierarchical channel-squeezing fusion layer to effectively extract relevant information from neighboring nodes while suppressing undesired noises in GNN learning. In addition, we propose a temporal-aware dynamic graph construction procedure that is robust and effective for 3D pose estimation. Experimental results on the Human3.6M dataset show that our solution achieves 10.3\% average prediction accuracy improvement and greatly improves on hard poses over state-of-the-art techniques. We further apply the proposed technique on the skeleton-based action recognition task and also achieve state-of-the-art performance. Our code is available at https://github.com/ailingzengzzz/Skeletal-GNN.

PDF Abstract ICCV 2021 PDF ICCV 2021 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
3D Human Pose Estimation Human3.6M Skeletal GNN (GT) Average MPJPE (mm) 30.4 # 34
Using 2D ground-truth joints Yes # 2
Multi-View or Monocular Monocular # 1
3D Human Pose Estimation Human3.6M Skeletal GNN Average MPJPE (mm) 47.9 # 137
Using 2D ground-truth joints No # 2
Multi-View or Monocular Monocular # 1
3D Human Pose Estimation MPI-INF-3DHP Skeletal GNN AUC 46.2 # 51
PCK 82.1 # 54
Skeleton Based Action Recognition NTU RGB+D Skeletal GNN Accuracy (CV) 96.7 # 19
Accuracy (CS) 91.6 # 25
Ensembled Modalities 4 # 2
Skeleton Based Action Recognition NTU RGB+D 120 Skeletal GNN Accuracy (Cross-Subject) 87.5 # 23
Accuracy (Cross-Setup) 89.2 # 21
Ensembled Modalities 4 # 1

Methods


No methods listed for this paper. Add relevant methods here