GazeTrack: High-Precision Eye Tracking Based on Regularization and Spatial Computing
By: Xiaoyin Yang
Potential Business Impact:
Makes virtual reality eyes track more accurately.
Eye tracking has become increasingly important in virtual and augmented reality applications; however, the current gaze accuracy falls short of meeting the requirements for spatial computing. We designed a gaze collection framework and utilized high-precision equipment to gather the first precise benchmark dataset, GazeTrack, encompassing diverse ethnicities, ages, and visual acuity conditions for pupil localization and gaze tracking. We propose a novel shape error regularization method to constrain pupil ellipse fitting and train on open-source datasets, enhancing semantic segmentation and pupil position prediction accuracy. Additionally, we invent a novel coordinate transformation method similar to paper unfolding to accurately predict gaze vectors on the GazeTrack dataset. Finally, we built a gaze vector generation model that achieves reduced gaze angle error with lower computational complexity compared to other methods.
Similar Papers
Eyes on Target: Gaze-Aware Object Detection in Egocentric Video
CV and Pattern Recognition
Helps computers see what people are looking at.
TimeGazer: Temporal Modeling of Predictive Gaze Stabilization for AR Interaction
Human-Computer Interaction
Makes AR games track your eyes better.
A deep learning approach to track eye movements based on events
CV and Pattern Recognition
Tracks eyes cheaply for better VR/AR games.