Improved itracker combined with bidirectional long short-term memory for 3D gaze estimation using appearance cues

Xiaolong Zhou, Jianing Lin, Zhuo Zhang, Zhanpeng Shao, Shenyong Chen, Honghai Liu

Research output: Contribution to journalArticlepeer-review

266 Downloads (Pure)

Abstract

Gaze is an important non-verbal cue for speculating human’s attention, which has been widely employed in many human–computer interaction-based applications. In this paper, we propose an improved Itracker to predict the subject’s gaze for a single image frame, as well as employ a many-to-one bidirectional Long Short-Term Memory (bi-LSTM) to fit the temporal information between frames to estimate gaze for video sequence. For single image frame gaze estimation, we improve the conventional Itracker by removing the face-grid and reducing one network branch via concatenating the two-eye region images. Experimental results show that our improved Itracker obtains 11.6% significant improvement over the state-of-the-art methods on MPIIGaze dataset and has robust estimation accuracy for different image resolutions under the premise of greatly reducing network complexity. For video sequence gaze estimation, by employing the bi-LSTM to fit the temporal information between frames, experimental results on EyeDiap dataset further demonstrate 3% accuracy improvement.
Original languageEnglish
JournalNeurocomputing
Early online date20 Oct 2019
DOIs
Publication statusEarly online - 20 Oct 2019

Fingerprint

Dive into the research topics of 'Improved itracker combined with bidirectional long short-term memory for 3D gaze estimation using appearance cues'. Together they form a unique fingerprint.

Cite this