4.6 Article

Predicting user visual attention in virtual reality with a deep learning model

期刊

VIRTUAL REALITY
卷 25, 期 4, 页码 1123-1136

出版社

SPRINGER LONDON LTD
DOI: 10.1007/s10055-021-00512-7

关键词

Visual attention; Virtual reality; Deep learning model; Eye tracking

资金

  1. Natural Science Foundation of China [61802341]
  2. ZJU-SUTD IDEA programme [IDEA006]

向作者/读者索取更多资源

Recent studies have shown that user's visual attention during virtual reality museum navigation can be estimated effectively with deep learning models. The ALRF model, which combines temporal-spatial features of user's eye movements and virtual object coordinates, outperformed state-of-the-art models with a prediction accuracy of 91.03%, demonstrating flexibility across different virtual reality environments.
Recent studies show that user's visual attention during virtual reality museum navigation can be effectively estimated with deep learning models. However, these models rely on large-scale datasets that usually are of high structure complexity and context specific, which is challenging for nonspecialist researchers and designers. Therefore, we present the deep learning model, ALRF, to generalise on real-time user visual attention prediction in virtual reality context. The model combines two parallel deep learning streams to process the compact dataset of temporal-spatial salient features of user's eye movements and virtual object coordinates. The prediction accuracy outperformed the state-of-the-art deep learning models by reaching record high 91.03%. Importantly, with quick parametric tuning, the model showed flexible applicability across different environments of the virtual reality museum and outdoor scenes. Implications for how the proposed model may be implemented as a generalising tool for adaptive virtual reality application design and evaluation are discussed.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据