Eye tracking has become increasingly important in virtual and augmented reality applications; however, the current gaze accuracy falls short of meeting the requirements for spatial computing. We designed a gaze collection framework and utilized high-precision equipment to gather the first precise benchmark dataset, GazeTrack, encompassing diverse ethnicities, ages, and visual acuity conditions for pupil localization and gaze tracking. We propose a novel shape error regularization method to constrain pupil ellipse fitting and train on open-source datasets, enhancing semantic segmentation and pupil position prediction accuracy. Additionally, we invent a novel coordinate transformation method similar to paper unfolding to accurately predict gaze vectors on the GazeTrack dataset. Finally, we built a gaze vector generation model that achieves reduced gaze angle error with lower computational complexity compared to other methods.
翻译:眼动追踪在虚拟现实与增强现实应用中日益重要,然而当前注视精度尚无法满足空间计算的需求。我们设计了一套注视数据采集框架,利用高精度设备收集了首个涵盖多种族、年龄及视力条件的精确基准数据集GazeTrack,用于瞳孔定位与注视追踪。我们提出了一种新颖的形状误差正则化方法,以约束瞳孔椭圆拟合,并在开源数据集上进行训练,从而提升语义分割与瞳孔位置预测的准确性。此外,我们发明了一种类似于纸张展开的新型坐标变换方法,能够在GazeTrack数据集上精确预测注视向量。最终,我们构建了一个注视向量生成模型,与其他方法相比,该模型在降低计算复杂度的同时实现了更小的注视角度误差。