4.7 Article

Accurate and efficient 3D hand pose regression for robot hand teleoperation using a monocular RGB camera

Journal

EXPERT SYSTEMS WITH APPLICATIONS
Volume 136, Issue -, Pages 327-337

Publisher

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.eswa.2019.06.055

Keywords

Hand pose estimation; Deep learning; Robot teleoperation; Monocular

Funding

  1. Spanish Government [TIN2016-76515R]
  2. Feder funds
  3. University of Alicante [GRE16-19]
  4. Valencian Government [GV/2018/022]
  5. Spanish grant for PhD studies [ACIF/2017/243]

Ask authors/readers for more resources

In this paper, we present a novel deep learning-based architecture, which is under the scope of expert and intelligent systems, to perform accurate real-time tridimensional hand pose estimation using a single RGB frame as an input, so there is no need to use multiple cameras or points of view, or RGB-D devices. The proposed pipeline is composed of two convolutional neural network architectures. The first one is in charge of detecting the hand in the image. The second one is able to accurately infer the tridimensional position of the joints retrieving, thus, the full hand pose. To do this, we captured our own large-scale dataset composed of images of hands and the corresponding 3D joints annotations. The proposal achieved a 3D hand pose mean error of below 5 mm on both the proposed dataset and Stereo Hand Pose Tracking Benchmark, which is a public dataset. Our method also outperforms the state-of-the-art methods. We also demonstrate in this paper the application of the proposal to perform a robotic hand teleoperation with high success. (C) 2019 Elsevier Ltd. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available