4.5 Article

Using efficient group pseudo-3D network to learn spatio-temporal features

期刊

SIGNAL IMAGE AND VIDEO PROCESSING
卷 15, 期 2, 页码 361-369

出版社

SPRINGER LONDON LTD
DOI: 10.1007/s11760-020-01758-5

关键词

Action classification; 3D convolutional network; Spatio-temporal features; Efficient network

资金

  1. National Natural Science Foundation of China [61772352]
  2. Science and Technology Planning Project of Sichuan Province [2019YFG0400, 2018GZDZX0031, 2018GZDZX0004, 2017GZDZX0003, 2018JY0182, 19ZDYF1286, 2020YFG0322]

向作者/读者索取更多资源

A study introduces an efficient group pseudo-3D (GP3D) convolution approach to reduce model size and computational power for action classification; By combining GP3D with MobileNetV3, the method saves parameters and maintains accuracy on the UCF-101 dataset; Compared to the inflated 3D convolutional network, GP3D achieves higher accuracy with significantly reduced model size.
Action classification is a challenging problem in computer vision in recent years; the three-dimensional convolutional neural network plays an important role in spatio-temporal feature extraction. However, the 3D convolution approach needs expensive computation and memory resources. This paper proposes an efficient group pseudo-3D (GP3D) convolution to reduce the model size and need less computational power. We built the GP3D with MobileNetV3 to extend the 2D pre-training parameters directly to the 3D convolutional network. We also used GP3D to replace the original inflated 3D convolutional network to efficiently reduce the model size. Compared with other state-of-the-art 3D convolutional networks, GP3D with the efficient network of MobileNetV3 can save about 3 to 22 times of parameters but maintain the same accuracy on the dataset of UCF-101. GP3D with an inflated 3D convolutional network can achieve about 90% top1 accuracy, while the model size is only about half of the original inflated 3D convolutional network.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据