Abstract

The landing point of a saccade defines the new fixation region, the new region of interest. We asked whether it was possible to predict the saccade landing point early in this very fast eye movement. This work proposes a new algorithm based on LSTM networks and a fine-grained loss function for saccade landing point prediction in real-world scenarios. Predicting the landing point is a critical milestone toward reducing the problems caused by display-update latency in gaze-contingent systems that make real-time changes in the display based on eye tracking. Saccadic eye movements are some of the fastest human neuro-motor activities with angular velocities of up to 1,000°/s. We present a comprehensive analysis of the performance of our method using a database with almost 220,000 saccades from 75 participants captured during natural viewing of videos. We include a comparison with state-of-the-art saccade landing point prediction algorithms. The results obtained using our proposed method outperformed existing approaches with improvements of up to 50% error reduction. Finally, we analyzed some factors that affected prediction errors including duration, length, age, and user intrinsic characteristics.

Highlights

  • Gaze-contingent displays [1] have been used in video streaming [2], [3], robot-assisted surgery [4], human-computer interfaces for new virtual reality environments [5], reading research [6] and simulation of impaired vision [7], [8], among others

  • While the common perception is that saccades are ballistic eye movements that are not modified during flight and have a straight path, this is often not the case [11]

  • Each user contributed to the test set with a different number of samples

Read more

Summary

Introduction

Gaze-contingent displays [1] have been used in video streaming [2], [3], robot-assisted surgery [4], human-computer interfaces for new virtual reality environments [5], reading research [6] and simulation of impaired vision [7], [8], among others. These systems allow researchers to investigate a variety of visual phenomena, including eye movement guidance in reading, stability of vision, visual search strategies, and scene perception. Due to the unavoidable latency of gaze-contingent systems (that are intended to update displays according to gaze location), the gaze is no longer at the measured location by the time the display can be updated

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call