Improving articulated hand pose detection for static finger sign recognition in RGB-D images

  • PDF / 5,437,657 Bytes
  • 45 Pages / 439.642 x 666.49 pts Page_size
  • 56 Downloads / 195 Views

DOWNLOAD

REPORT


Improving articulated hand pose detection for static finger sign recognition in RGB-D images Abdessamad Elboushaki1

· Rachida Hannane2 · Karim Afdel2 · Lahcen Koutti2

Received: 15 May 2019 / Revised: 12 June 2020 / Accepted: 16 July 2020 / © Springer Science+Business Media, LLC, part of Springer Nature 2020

Abstract With the emergence of consumer RGB-D sensors, discriminative modeling has been shown to perform well in estimating human body pose. However, articulated hand pose estimation remains a challenging problem, mostly due to its high flexibility, occlusions, noisy data, and small area of the fingertips. In this paper, we present an efficient discriminative-based scheme to improve the performance of hand pose estimation from a single depth image. The proposed scheme is inspired by decision forest-based framework, but with several well-motivated modifications. Specifically, we propose a method to estimate 2D in-plane orientation of the hand, which is then utilized to enforce the depth comparison features and make them invariant to in-plane rotation. Subsequently, we investigate the use of random decision forests (RDF) and mean shift algorithm to predict a primary version of hand parts and joint locations. Based on this primary prediction, an adaptive spatial clustering method is applied to correct the misclassified regions, and to deliver the final estimation of hand pose. Along with the proposed scheme, we further develop a new set of highly-distinctive features for static finger sign recognition by utilizing the estimated hand pose configurations and RGB information. The proposed features are straightforward and can effectively capture different aspects of hand pose, such as links from each joint to the closest joints and orientation of each hand part. Extensive experiments on several challenging datasets demonstrate that our approach, compared to decision forest-based methods, is able to provide more precise estimation of hand poses (with up to 21% improvement in joint localization accuracy), and can efficiently recognize more complex static finger signs (93.85% mean recognition accuracy on a challenging 34-finger sign dataset). Our approach is also robust to illumination, inter-hand occlusion, scale, and rotation variance. Keywords Hand pose estimation · Finger sign recognition · Per-pixel classification · Random decision forest · Density based spatial clustering · RGB-D information

 Abdessamad Elboushaki

[email protected]; [email protected]

Extended author information available on the last page of the article.

Multimedia Tools and Applications

1 Introduction The use of vision-based hand representation in Human-Computer Interaction (HCI) has been received great interests in recent years due to its wide range of successful applications in pattern recognition and computer vision fields, including: computer games [76], robot control [23], surgical and medical assistive systems [46], user interface controls [8], and many more. Among potential research areas, hand pose esti