Pixel-Level Hand Detection in Ego-centric Videos

Cheng Li, Kris M. Kitani; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2013, pp. 3570-3577

Abstract


We address the task of pixel-level hand detection in the context of ego-centric cameras. Extracting hand regions in ego-centric videos is a critical step for understanding handobject manipulation and analyzing hand-eye coordination. However, in contrast to traditional applications of hand detection, such as gesture interfaces or sign-language recognition, ego-centric videos present new challenges such as rapid changes in illuminations, significant camera motion and complex hand-object manipulations. To quantify the challenges and performance in this new domain, we present a fully labeled indoor/outdoor ego-centric hand detection benchmark dataset containing over 200 million labeled pixels, which contains hand images taken under various illumination conditions. Using both our dataset and a publicly available ego-centric indoors dataset, we give extensive analysis of detection performance using a wide range of local appearance features. Our analysis highlights the effectiveness of sparse features and the importance of modeling global illumination. We propose a modeling strategy based on our findings and show that our model outperforms several baseline approaches.

Related Material


[pdf]
[bibtex]
@InProceedings{Li_2013_CVPR,
author = {Li, Cheng and Kitani, Kris M.},
title = {Pixel-Level Hand Detection in Ego-centric Videos},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2013}
}