CrowdPose: Efficient Crowded Scenes Pose Estimation and A New Benchmark

Multi-person pose estimation is fundamental to many computer vision tasks and has made significant progress in recent years. However, few previous methods explored the problem of pose estimation in crowded scenes while it remains challenging and inevitable in many scenarios. Moreover, current benchmarks cannot provide an appropriate evaluation for such cases. In this paper, we propose a novel and efficient method to tackle the problem of pose estimation in the crowd and a new dataset to better evaluate algorithms. Our model consists of two key components: joint-candidate single person pose estimation (SPPE) and global maximum joints association. With multi-peak prediction for each joint and global association using graph model, our method is robust to inevitable interference in crowded scenes and very efficient in inference. The proposed method surpasses the state-of-the-art methods on CrowdPose dataset by 5.2 mAP and results on MSCOCO dataset demonstrate the generalization ability of our method. Source code and dataset will be made publicly available.

PDF Abstract CVPR 2019 PDF CVPR 2019 Abstract

Datasets


Introduced in the Paper:

CrowdPose

Used in the Paper:

MS COCO MPII OCHuman
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Multi-Person Pose Estimation CrowdPose Joint-candidate SPPE + mAP @0.5:0.95 66.0 # 13
AP Easy 75.5 # 11
AP Medium 66.3 # 13
AP Hard 57.4 # 11
FPS 10.1 # 3

Results from Other Papers


Task Dataset Model Metric Name Metric Value Rank Source Paper Compare
Multi-Person Pose Estimation OCHuman CrowdPose Validation AP 27.5 # 5
AP50 40.8 # 6
AP75 29.9 # 6

Methods


No methods listed for this paper. Add relevant methods here