4.6 Article

Fusing two-stream convolutional neural networks for RGB-T object tracking

期刊

NEUROCOMPUTING
卷 281, 期 -, 页码 78-85

出版社

ELSEVIER
DOI: 10.1016/j.neucom.2017.11.068

关键词

Object tracking; Thermal information; Adaptive fusion; Convolutional neural network; Correlation filter

资金

  1. National Natural Science Foundation of China [61702002, 61472002, 61602006]
  2. Natural Science Foundation of Anhui Higher Education Institution of China [KJ2017A017, KJ2015A110]
  3. Co-Innovation Center for Information Supply AMP
  4. Assurance Technology, Anhui University [Y01002449]

向作者/读者索取更多资源

This paper investigates how to integrate the complementary information from RGB and thermal (RGB-T) sources for object tracking. We propose a novel Convolutional Neural Network (ConvNet) architecture, including a two-stream ConvNet and a FusionNet, to achieve adaptive fusion of different source data for robust RGB-T tracking. Both RGB and thermal streams extract generic semantic information of the target object. In particular, the thermal stream is pre-trained on the ImageNet dataset to encode rich semantic information, and then fine-tuned using thermal images to capture the specific properties of thermal information. For adaptive fusion of different modalities while avoiding redundant noises, the FusionNet is employed to select most discriminative feature maps from the outputs of the two-stream ConvNet, and updated online to adapt to appearance variations of the target object. Finally, the object locations are efficiently predicted by applying the multi-channel correlation filter on the fused feature maps. Extensive experiments on the recently public benchmark GTOT verify the effectiveness of the proposed approach against other state-of-the-art RGB-T trackers. (c) 2017 Published by Elsevier B.V.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据