Please wait a minute...
IMAGE/TABLE DETAILS
Extracting hand articulations from monocular depth images using curvature scale space descriptors
Shao-fan WANG,Chun LI,De-hui KONG,Bao-cai YIN
Front. Inform. Technol. Electron. Eng.    2016, 17 (1): 41-54.   DOI: 10.1631/FITEE.1500126
Abstract   HTML   PDF (2840KB)

We propose a framework of hand articulation detection from a monocular depth image using curvature scale space (CSS) descriptors. We extract the hand contour from an input depth image, and obtain the fingertips and finger-valleys of the contour using the local extrema of a modified CSS map of the contour. Then we recover the undetected fingertips according to the local change of depths of points in the interior of the contour. Compared with traditional appearance-based approaches using either angle detectors or convex hull detectors, the modified CSS descriptor extracts the fingertips and finger-valleys more precisely since it is more robust to noisy or corrupted data; moreover, the local extrema of depths recover the fingertips of bending fingers well while traditional appearance-based approaches hardly work without matching models of hands. Experimental results show that our method captures the hand articulations more precisely compared with three state-of-the-art appearance-based approaches.




View image in article
Fig. 12 Failure examples of the CSS method of Experiment 2 (red: fingertips; blue: finger-valleys; yellow: palm center). References to color refer to the online version of this figure
Extracts from the Article
Fig. 12 shows some failure examples of Experiment 2 using the CSS method. We consider those examples as failures since our method misses or incorrectly detects at least two fingertips of the target hand. In general, our method fails for the examples with large occlusions or with small resolutions. Such a disadvantage is common for appearance-based methods and can be improved by model-based methods. This is the future work we shall consider.
Other Images/Table from this Article