4.7 Article

Web Video Event Recognition by Semantic Analysis From Ubiquitous Documents

期刊

IEEE TRANSACTIONS ON IMAGE PROCESSING
卷 26, 期 12, 页码 5689-5701

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2016.2614136

关键词

Video event recognition; event knowledge base; two-view adaptive regression

资金

  1. Australian Research Council [DP130103252, FT120100718]
  2. National Natural Science Foundation of China [61572108]
  3. National Thousand-Young-Talents Program of China
  4. Fundamental Research Funds for the Central Universities [ZYGX2014Z007, ZYGX2015J055]
  5. Australian Research Council [FT120100718] Funding Source: Australian Research Council

向作者/读者索取更多资源

In recent years, the task of event recognition from videos has attracted increasing interest in multimedia area. While most of the existing research was mainly focused on exploring visual cues to handle relatively small-granular events, it is difficult to directly analyze video content without any prior knowledge. Therefore, synthesizing both the visual and semantic analysis is a natural way for video event understanding. In this paper, we study the problem of Web video event recognition, where Web videos often describe large-granular events and carry limited textual information. Key challenges include how to accurately represent event semantics from incomplete textual information and how to effectively explore the correlation between visual and textual cues for video event understanding. We propose a novel framework to perform complex event recognition from Web videos. In order to compensate the insufficient expressive power of visual cues, we construct an event knowledge base by deeply mining semantic information from ubiquitous Web documents. This event knowledge base is capable of describing each event with comprehensive semantics. By utilizing this base, the textual cues for a video can be significantly enriched. Furthermore, we introduce a two-view adaptive regression model, which explores the intrinsic correlation between the visual and textual cues of the videos to learn reliable classifiers. Extensive experiments on two real-world video data sets show the effectiveness of our proposed framework and prove that the event knowledge base indeed helps improve the performance of Web video event recognition.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据