Estimating Gaze From Head And Hand Pose And Scene Images For Open-Ended Exploration In Vr Environments

2021 IEEE CONFERENCE ON VIRTUAL REALITY AND 3D USER INTERFACES ABSTRACTS AND WORKSHOPS (VRW 2021)(2021)

引用 8|浏览53
暂无评分
摘要
The widespread utility of eye tracking technology has created a growing demand for more consistent and reliable eye-tracking systems, and there is a need for new and accessible approaches that can enhance the accuracy of eye-tracking data. Previous studies have offered evidence for associations between certain non-eye signals and gaze such as a strong coordination between head motion and gaze shifts. e.g. [3] , hand and eye spatiotemporal statistics, e.g. [7] , and gaze behavior and scene content, e.g. [2] . Previous studies have also shown how various combinations of eye, head, scene, and hand signals can be leveraged for applications such as gaze estimation [5] , [10] , prediction [8] , and classification [6] . Though these previous approaches provide support for the idea that non-eye sensors (i.e. head, hand, and scene) are useful for estimating gaze, they have not yet fully addressed how these signals individually and in combination contribute to gaze estimation.
更多
查看译文
关键词
eye tracking,gaze estimation,virtual reality,non eye sensors
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要