Unconstrained Monocular 3D Human Pose Estimation by Action Detection and Cross-Modality Regression Forest

Tsz-Ho Yu, Tae-Kyun Kim, Roberto Cipolla; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2013, pp. 3642-3649

Abstract


This work addresses the challenging problem of unconstrained 3D human pose estimation (HPE) from a novel perspective. Existing approaches struggle to operate in realistic applications, mainly due to their scene-dependent priors, such as background segmentation and multi-camera network, which restrict their use in unconstrained environments. We therfore present a framework which applies action detection and 2D pose estimation techniques to infer 3D poses in an unconstrained video. Action detection offers spatiotemporal priors to 3D human pose estimation by both recognising and localising actions in space-time. Instead of holistic features, e.g. silhouettes, we leverage the flexibility of deformable part model to detect 2D body parts as a feature to estimate 3D poses. A new unconstrained pose dataset has been collected to justify the feasibility of our method, which demonstrated promising results, significantly outperforming the relevant state-of-the-arts.

Related Material


[pdf]
[bibtex]
@InProceedings{Yu_2013_CVPR,
author = {Yu, Tsz-Ho and Kim, Tae-Kyun and Cipolla, Roberto},
title = {Unconstrained Monocular 3D Human Pose Estimation by Action Detection and Cross-Modality Regression Forest},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2013}
}