Abstract

Detecting the pupil center plays a key role in human-computer interaction, especially for gaze tracking. The conventional deep learning-based method for this problem is to train a convolutional neural network (CNN), which takes the eye image as the input and gives the pupil center as a regression result. In this paper, we propose an indirect use of the CNN for the task, which first segments the pupil region by a CNN as a classification problem, and then finds the center of the segmented region. This is based on the observation that CNN works more robustly for the pupil segmentation than for the pupil center-point regression when the inputs are noisy IR images. Specifically, we use the UNet model for the segmentation of pupil regions in IR images and then find the pupil center as the center of mass of the segment. In designing the loss function for the segmentation, we propose a new loss term that encodes the convex shape-prior for enhancing the robustness to noise. Precisely, we penalize not only the deviation of each predicted pixel from the ground truth label but also the non-convex shape of pupils caused by the noise and reflection. For the training, we make a new dataset of 111,581 images with hand-labeled pupil regions from 29 IR eye video sequences. We also label commonly used datasets ( ExCuSe and ElSe dataset) that are considered real-world noisy ones to validate our method. Experiments show that the proposed method performs better than the conventional methods that directly find the pupil center as a regression result.

Highlights

  • Eye tracking or gaze tracking is one of the most important techniques for Human-Computer Interaction (HCI) and its applications

  • 2) We propose a new loss term that encodes convex shape prior for the training of segmentation networks, which further increases the accuracy of pupil center estimates

  • We find the center of the largest blob as a pupil center

Read more

Summary

Introduction

Eye tracking or gaze tracking is one of the most important techniques for Human-Computer Interaction (HCI) and its applications It is essential for pointing (virtual) objects in AR/VR environment [1]–[3], detecting drowsiness that improves driver safety [4], [5], analyzing the human behavior such as eye-tracking heat map [6], [7], etc. These methods usually need real-time eye-tracking, for interacting with virtual objects or for the foveated image rendering in VR environments [8].

Methods
Results
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.