3D Multi-Person Pose Estimation
18 papers with code • 5 benchmarks • 4 datasets
This task aims to solve root-relative 3D multi-person pose estimation. No human bounding box and root joint coordinate groundtruth are used in testing time.
( Image credit: RootNet )
Libraries
Use these libraries to find 3D Multi-Person Pose Estimation models and implementationsSubtasks
Most implemented papers
End-to-end Recovery of Human Shape and Pose
The main objective is to minimize the reprojection loss of keypoints, which allow our model to be trained using images in-the-wild that only have ground truth 2D annotations.
Camera Distance-aware Top-down Approach for 3D Multi-person Pose Estimation from a Single RGB Image
Although significant improvement has been achieved recently in 3D human pose estimation, most of the previous methods only treat a single-person case.
Fast and Robust Multi-Person 3D Pose Estimation from Multiple Views
This paper addresses the problem of 3D pose estimation for multiple people in a few calibrated camera views.
VoxelPose: Towards Multi-Camera 3D Human Pose Estimation in Wild Environment
In contrast to the previous efforts which require to establish cross-view correspondence based on noisy and incomplete 2D pose estimations, we present an end-to-end solution which directly operates in the $3$D space, therefore avoids making incorrect decisions in the 2D space.
Learning to Reconstruct 3D Human Pose and Shape via Model-fitting in the Loop
Our approach is self-improving by nature, since better network estimates can lead the optimization to better solutions, while more accurate optimization fits provide better supervision for the network.
4D Association Graph for Realtime Multi-person Motion Capture Using Multiple Video Cameras
Our method enables a realtime online motion capture system running at 30fps using 5 cameras on a 5-person scene.
Cross-View Tracking for Multi-Human 3D Pose Estimation at over 100 FPS
To further verify the scalability of our method, we propose a new large-scale multi-human dataset with 12 to 28 camera views.
Unsupervised Cross-Modal Alignment for Multi-Person 3D Pose Estimation
Our approach not only generalizes to in-the-wild images, but also yields a superior trade-off between speed and performance, compared to prior top-down approaches.
Monocular, One-stage, Regression of Multiple 3D People
Through a body-center-guided sampling process, the body mesh parameters of all people in the image are easily extracted from the Mesh Parameter map.
Temporal Smoothing for 3D Human Pose Estimation and Localization for Occluded People
In multi-person pose estimation actors can be heavily occluded, even become fully invisible behind another person.