4.7 Article

Learning the Relative Importance of Objects from Tagged Images for Retrieval and Cross-Modal Search

期刊

INTERNATIONAL JOURNAL OF COMPUTER VISION
卷 100, 期 2, 页码 134-153

出版社

SPRINGER
DOI: 10.1007/s11263-011-0494-3

关键词

Image retrieval; Image tags; Multi-modal retrieval; Cross-modal retrieval; Image search; Object recognition; Auto annotation; Kernelized canonical correlation analysis

资金

  1. Luce Foundation
  2. DARPA CSSG [N11AP20004]

向作者/读者索取更多资源

We introduce an approach to image retrieval and auto-tagging that leverages the implicit information about object importance conveyed by the list of keyword tags a person supplies for an image. We propose an unsupervised learning procedure based on Kernel Canonical Correlation Analysis that discovers the relationship between how humans tag images (e.g., the order in which words are mentioned) and the relative importance of objects and their layout in the scene. Using this discovered connection, we show how to boost accuracy for novel queries, such that the search results better preserve the aspects a human may find most worth mentioning. We evaluate our approach on three datasets using either keyword tags or natural language descriptions, and quantify results with both ground truth parameters as well as direct tests with human subjects. Our results show clear improvements over approaches that either rely on image features alone, or that use words and image features but ignore the implied importance cues. Overall, our work provides a novel way to incorporate high-level human perception of scenes into visual representations for enhanced image search.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据