Jiang Wang, Zicheng Liu, Jan Chorowski, Zhuoyuan Chen, and Ying Wu
7 October 2012
We study the problem of action recognition from depth sequences
captured by depth cameras, where noise and occlusion are common
problems because they are captured with a single commodity camera.
In order to deal with these issues, we extract semi-local features
called random occupancy pattern (ROP) features, which employ a novel
sampling scheme that effectively explores an extremely large sampling
space. We also utilize a sparse coding approach to robustly encode these
features. The proposed approach does not require careful parameter tuning.
Its training is very fast due to the use of the high-dimensional integral
image, and it is robust to the occlusions. Our technique is evaluated on
two datasets captured by commodity depth cameras: an action dataset
and a hand gesture dataset. Our classification results are superior to
those obtained by the state of the art approaches on both datasets.
|Published in||12th European Conference on Computer Vision (ECCV)|
© 2012 Springer. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the Springer.