4.6 Article

Fusion of 2D CNN and 3D DenseNet for Dynamic Gesture Recognition

Journal

ELECTRONICS
Volume 8, Issue 12, Pages -

Publisher

MDPI
DOI: 10.3390/electronics8121511

Keywords

gesture recognition; motion representation; 2D CNN; 3D DenseNet; information fusion

Funding

  1. Key Program of Natural Science Foundation of Shaanxi Province of China [2017JZ020]
  2. National Natural Science Foundation of China [61771386, 61671374]

Ask authors/readers for more resources

Gesture recognition has been applied in many fields as it is a natural human-computer communication method. However, recognition of dynamic gesture is still a challenging topic because of complex disturbance information and motion information. In this paper, we propose an effective dynamic gesture recognition method by fusing the prediction results of a two-dimensional (2D) motion representation convolution neural network (CNN) model and three-dimensional (3D) dense convolutional network (DenseNet) model. Firstly, to obtain a compact and discriminative gesture motion representation, the motion history image (MHI) and pseudo-coloring technique were employed to integrate the spatiotemporal motion sequences into a frame image, before being fed into a 2D CNN model for gesture classification. Next, the proposed 3D DenseNet model was used to extract spatiotemporal features directly from Red, Green, Blue (RGB) gesture videos. Finally, the prediction results of the proposed 2D and 3D deep models were blended together to boost recognition performance. The experimental results on two public datasets demonstrate the effectiveness of our proposed method.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available