Capturing Humans in Motion: Temporal-Attentive 3D Human Pose and Shape Estimation from Monocular Video

Learning to capture human motion is essential to 3D human pose and shape estimation from monocular video. However, the existing methods mainly rely on recurrent or convolutional operation to model such temporal information, which limits the ability to capture non-local context relations of human motion. To address this problem, we propose a motion pose and shape network (MPS-Net) to effectively capture humans in motion to estimate accurate and temporally coherent 3D human pose and shape from a video. Specifically, we first propose a motion continuity attention (MoCA) module that leverages visual cues observed from human motion to adaptively recalibrate the range that needs attention in the sequence to better capture the motion continuity dependencies. Then, we develop a hierarchical attentive feature integration (HAFI) module to effectively combine adjacent past and future feature representations to strengthen temporal correlation and refine the feature representation of the current frame. By coupling the MoCA and HAFI modules, the proposed MPS-Net excels in estimating 3D human pose and shape in the video. Though conceptually simple, our MPS-Net not only outperforms the state-of-the-art methods on the 3DPW, MPI-INF-3DHP, and Human3.6M benchmark datasets, but also uses fewer network parameters. The video demos can be found at https://mps-net.github.io/MPS-Net/.

PDF Abstract CVPR 2022 PDF CVPR 2022 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
3D Human Pose Estimation 3DPW MPS-Net (T=16) PA-MPJPE 52.1 # 67
MPJPE 84.3 # 69
MPVPE 99.7 # 52
Acceleration Error 7.4 # 7
Number of parameters (M) 39.63 # 1
FLOPs (G) 4.45 # 1
3D Human Pose Estimation Human3.6M MPS-Net (T=16) Average MPJPE (mm) 69.4 # 276
PA-MPJPE 47.4 # 87
Acceleration Error 3.6 # 3
3D Human Pose Estimation MPI-INF-3DHP MPS-Net (T=16) MPJPE 96.7 # 59
PA-MPJPE 62.8 # 11
Acceleration Error 9.6 # 7

Methods


No methods listed for this paper. Add relevant methods here