INTAIRACT: Joint hand gesture and fingertip classification for touchless interaction
Document typeConference lecture
Rights accessRestricted access - publisher's policy
In this demo we present intAIRact, an online hand-based touchless interaction system. Interactions are based on easy-to-learn hand gestures, that combined with translations and rotations render a user friendly and highly configurable system. The main advantage with respect to existing approaches is that we are able to robustly locate and identify fingertips. Hence, we are able to employ a simple but powerful alphabet of gestures not only by determining the number of visible fingers in a gesture, but also which fingers are being observed. To achieve such a system we propose a novel method that jointly infers hand gestures and fingertip locations using a single depth image from a consumer depth camera. Our approach is based on a novel descriptor for depth data, the Oriented Radial Distribution (ORD) . On the one hand, we exploit the ORD for robust classification of hand gestures by means of efficient k-NN retrieval. On the other hand, maxima of the ORD are used to perform structured inference of fingertip locations. The proposed method outperforms other state-of-the-art approaches both in gesture recognition and fingertip localization. An implementation of the ORD extraction on a GPU yields a real-time demo running at approximately 17fps on a single laptop
CitationSuau, X. [et al.]. INTAIRACT: Joint hand gesture and fingertip classification for touchless interaction. A: European Conference on Computer Vision. "Computer Vision – ECCV 2012. Workshops and Demonstrations". Florència: Springer, 2012, p. 602-606.
|INTAIRACTJointH ... orTouchlessInteraction.pdf||INTAIRACT: Joint Hand Gesture and Fingertip Classification for Touchless Interaction||203,0Kb||Restricted access|