4.6 Article

Time-varying LSTM networks for action recognition

Journal

MULTIMEDIA TOOLS AND APPLICATIONS
Volume 77, Issue 24, Pages 32275-32285

Publisher

SPRINGER
DOI: 10.1007/s11042-018-6260-6

Keywords

RNNs; CNNs; LSTMs; TV-LSTMs; Action recognition

Funding

  1. National Natural Science Foundation of China [61672299]

Ask authors/readers for more resources

We describe an architecture of Time-Varying Long Short-Term Memory recurrent neural networks (TV-LSTMs) for human action recognition. The main innovation of this architecture is the use of hybrid weights, shared weights and non-shared weights which we refer to as varying weights. The varying weights can enhance the ability of LSTMs to represent videos and other sequential data. We evaluate TV-LSTMs on UCF-11, HMDB-51, and UCF-101 human action datasets and achieve the top-1 accuracy of 99.64%, 57.52%, and 85.06% respectively. This model performs competitively against the models that use both RGB and other features, such as optical flows, improved Dense Trajectory, etc. In this paper, we also propose and analyze the methods of selecting varying weights.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available