Abstract
Appearance-based methods with deep learning can predict the point of gaze by using a monocular camera, which requires a large sample to learn. However, existing appearance-based gaze estimation methods with deep learning mainly use face and eye images or only use a single face image, ignoring the correlation between face features and eye features In response to this issue, we propose a coordination model where face feature extraction is the gaze estimation network and eye feature extraction is the coordination network, which deeply fuses the eye-face feature relationships to perform the gaze estimation task. The model achieves good results on MPIIFaceGaze dataset and GazeCapture dataset.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.