Journal
MULTIMEDIA TOOLS AND APPLICATIONS
Volume 77, Issue 24, Pages 32275-32285Publisher
SPRINGER
DOI: 10.1007/s11042-018-6260-6
Keywords
RNNs; CNNs; LSTMs; TV-LSTMs; Action recognition
Categories
Funding
- National Natural Science Foundation of China [61672299]
Ask authors/readers for more resources
We describe an architecture of Time-Varying Long Short-Term Memory recurrent neural networks (TV-LSTMs) for human action recognition. The main innovation of this architecture is the use of hybrid weights, shared weights and non-shared weights which we refer to as varying weights. The varying weights can enhance the ability of LSTMs to represent videos and other sequential data. We evaluate TV-LSTMs on UCF-11, HMDB-51, and UCF-101 human action datasets and achieve the top-1 accuracy of 99.64%, 57.52%, and 85.06% respectively. This model performs competitively against the models that use both RGB and other features, such as optical flows, improved Dense Trajectory, etc. In this paper, we also propose and analyze the methods of selecting varying weights.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available