4.6 Article

Integrating Scene Semantic Knowledge into Image Captioning

Publisher

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3439734

Keywords

Image captioning; attention mechanism; scene semantics; encoder-decoder framework

Funding

  1. National Natural Science Foundation of China [61966004, 61663004, 61866004, 61762078]
  2. Guangxi Natural Science Foundation [2019GXNSFDA245018, 2018GXNSFDA281009]
  3. Guangxi Bagui Scholar Teams for Innovation and Research Project
  4. Guangxi Talent Highland Project of Big Data Intelligence and Application
  5. Guangxi Collaborative Innovation Center of Multi-Source Information Integration and Intelligent Processing

Ask authors/readers for more resources

Most existing image captioning methods lack effective scene semantic information and cannot adjust focus intensity on the image. This article proposes an improved visual attention model that combines visual and semantic information to generate more accurate, scene-specific captions.
Most existing image captioning methods use only the visual information of the image to guide the generation of captions, lack the guidance of effective scene semantic information, and the current visual attention mechanism cannot adjust the focus intensity on the image. In this article, we first propose an improved visual attention model. At each timestep, we calculated the focus intensity coefficient of the attention mechanism through the context information of themodel, then automatically adjusted the focus intensity of the attention mechanism through the coefficient to extract more accurate visual information. In addition, we represented the scene semantic knowledge of the image through topic words related to the image scene, then added them to the language model. We used the attention mechanism to determine the visual information and scene semantic information that the model pays attention to at each timestep and combined them to enable the model to generate more accurate and scene-specific captions. Finally, we evaluated our model on Microsoft COCO (MSCOCO) and Flickr30k standard datasets. The experimental results show that our approach generates more accurate captions and outperforms many recent advanced models in various evaluation metrics.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available