Gazeformer: Scalable, Effective and Fast Prediction of Goal-Directed Human Attention

March 2, 2023

Predicting human gaze is important in Human-Computer Interaction (HCI). However, to practically serve HCI applications, gaze prediction models must be scalable, fast, and accurate in their spatial and temporal gaze predictions. Recent scanpath prediction models focus on goal-directed attention (search). Such models are limited in their application due to a common approach relying on trained target detectors for all possible objects, and the availability of human gaze data for their training (both not scalable). In response, we pose a new task called ZeroGaze, a new variant of zero-shot learning where gaze is predicted for never-before-searched objects, and we develop a novel model, Gazeformer, to solve the ZeroGaze problem. In contrast to existing methods using object detector modules, Gazeformer encodes the target using a natural language model, thus leveraging semantic similarities in scanpath prediction. We use a transformer-based encoder-decoder architecture because transformers are particularly useful for generating contextual representations. Gazeformer surpasses other models by a large margin on the ZeroGaze setting. It also outperforms existing target-detection models on standard gaze prediction for both target-present and target-absent search tasks. In addition to its improved performance, Gazeformer is more than five times faster than the state-of-the-art target-present visual search model.


< 1 minute

Sounak Mondal, Zhibo Yang, Seoyoung Ahn, Dimitris Samaras, Greg Zelinsky, Minh Hoai

CVPR 2023

Share Article

Related publications

CV AAAI Top Tier
January 8, 2024

Yifeng*, Duc Nguyen Duy*, Lam Nguyen Thanh, Cuong Pham, Minh Hoai

January 8, 2024

Tran Huynh Ngoc, Dang Minh Nguyen, Tung Pham, Anh Tran

CV NeurIPS Top Tier
October 4, 2023

Quang Nguyen, Vu Tuan Truong, Anh Tran, Khoi Nguyen

CV NeurIPS Top Tier
October 4, 2023

Dung Nguyen, Tuan Nguyen, Anh Tran, Khoa Doan, Kok-seng Wong