4.7 Article

Global-prior-guided fusion network for salient object detection

期刊

EXPERT SYSTEMS WITH APPLICATIONS
卷 198, 期 -, 页码 -

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.eswa.2022.116805

关键词

Global prior; RGB-D salient object detection; Saliency detection; Global fusion strategy

资金

  1. Key Area R&D Program of Guang-dong Province, China [2018B030338001]
  2. National Natural Sci-ence Foundation of China [62076055, 61703075]
  3. 111 Project of China [B12027]

向作者/读者索取更多资源

In this paper, a novel global-prior-guided fusion network is proposed for RGB-D based salient object detection. By extracting global features and guiding the processing of local features, the method outperforms other state-of-the-art methods on multiple datasets and shows scalability across different tasks.
RGB image and depth map (RGB-D)-based salient object detection (SOD) has been well-studied in recent years, especially using deep neural networks. An RGB image provides rich local and semantic features, while the depth map provides global structural information. Many researchers have treated depth information as a supplement to RGB maps. However, depth maps in various datasets are not as precise as RGB information, as they are captured under various conditions. Therefore, thoroughly exploiting these features at different levels remains unresolved. Many cognitive theories, such as the topological perception theory, claim that global properties are prior to local ones and are important for human cognition. In this paper, we propose a novel global-prior-guided fusion network with global-prior extraction modules to fuse cross-modality features. Each module contains a cross attention guided by deeper global priors, and the global prior extracted by this module is used to guide the processing of local features in shallow layers. The global guided network first integrates the local and global cross features into the decoder of depth maps, and then the fused structural features of the decoder are finally fused into the saliency decoder. Experimental results show that our method outperformed other state-of-the-art methods in the RGB-D-based SOD task on seven datasets (i.e., DUT-RGBD, NJUD, LFSD, NLPR, RGBD135, SIP, and STERE) and in terms of most metrics. To thoroughly exploit the modules we designed, we extended our model to accomplish the tasks of RGB and video SOD with slight adaptions, and obtained results comparable to those of the state-of-the-art (SOTA) methods in both fields.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据