4.7 Article

Transferable Visual Words: Exploiting the Semantics of Anatomical Patterns for Self-Supervised Learning

期刊

IEEE TRANSACTIONS ON MEDICAL IMAGING
卷 40, 期 10, 页码 2857-2868

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TMI.2021.3060634

关键词

Visualization; Semantics; Image representation; Feature extraction; Biomedical imaging; Annotations; Training; Self-supervised learning; transfer learning; visual words; anatomical patterns; computational anatomy; 3D medical imaging; and 3D pre-trained models

资金

  1. ASU
  2. NIH [R01HL128785]
  3. GPUs through the ASU Research Computing
  4. Extreme Science and Engineering Discovery Environment (XSEDE) - National Science Foundation (NSF) [ACI-1548562]
  5. Mayo Clinic

向作者/读者索取更多资源

This paper introduces a new concept called TransVW, which aims to improve the annotation efficiency of deep learning in medical image analysis by automatically harvesting visual words and utilizing self-supervised learning. Experimental results demonstrate that TransVW offers higher performance and faster convergence while reducing annotation costs.
This paper introduces a new concept called transferable visual words (TransVW), aiming to achieve annotation efficiency for deep learning in medical image analysis. Medical imaging-focusing on particular parts of the body for defined clinical purposes-generates images of great similarity in anatomy across patients and yields sophisticated anatomical patterns across images, which are associated with rich semantics about human anatomy and which are natural visual words. We show that these visual words can be automatically harvested according to anatomical consistency via self-discovery, and that the self-discovered visual words can serve as strong yet free supervision signals for deep models to learn semantics-enriched generic image representation via self-supervision (self-classification and self-restoration). Our extensive experiments demonstrate the annotation efficiency of TransVW by offering higher performance and faster convergence with reduced annotation cost in several applications. Our TransVW has several important advantages, including (1) TransVW is a fully autodidactic scheme, which exploits the semantics of visual words for self-supervised learning, requiring no expert annotation; (2) visual word learning is an add-on strategy, which complements existing self-supervised methods, boosting their performance; and (3) the learned image representation is semantics-enriched models, which have proven to be more robust and generalizable, saving annotation efforts for a variety of applications through transfer learning. Our code, pre-trained models, and curated visual words are available at https://github.com/JLiangLab/TransVW.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据