4.5 Article

STATE: Learning structure and texture representations for novel view synthesis

期刊

COMPUTATIONAL VISUAL MEDIA
卷 9, 期 4, 页码 767-786

出版社

SPRINGERNATURE
DOI: 10.1007/s41095-022-0301-9

关键词

novel view synthesis; sparse views; spatio-view attention; structure representation; texture representation

向作者/读者索取更多资源

In this paper, an end-to-end deep neural network called STATE is proposed for sparse view synthesis by learning structure and texture representations. The network encodes structure as a hybrid feature field and texture as a deformed feature map. It employs a hierarchical fusion scheme with spatio-view attention to adaptively select important information. Experimental results show that the proposed method achieves better results than state-of-the-art methods in both qualitative and quantitative evaluations. The method also enables texture and structure editing applications benefiting from implicit disentanglement of structure and texture.
Novel viewpoint image synthesis is very challenging, especially from sparse views, due to large changes in viewpoint and occlusion. Existing image-based methods fail to generate reasonable results for invisible regions, while geometry-based methods have difficulties in synthesizing detailed textures. In this paper, we propose STATE, an end-to-end deep neural network, for sparse view synthesis by learning structure and texture representations. Structure is encoded as a hybrid feature field to predict reasonable structures for invisible regions while maintaining original structures for visible regions, and texture is encoded as a deformed feature map to preserve detailed textures. We propose a hierarchical fusion scheme with intra-branch and inter-branch aggregation, in which spatio-view attention allows multi-view fusion at the feature level to adaptively select important information by regressing pixel-wise or voxel-wise confidence maps. By decoding the aggregated features, STATE is able to generate realistic images with reasonable structures and detailed textures. Experimental results demonstrate that our method achieves qualitatively and quantitatively better results than state-of-the-art methods. Our method also enables texture and structure editing applications benefiting from implicit disentanglement of structure and texture.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据