Resource-efficient Gaze Estimation via Frequency-domain Multi-task Contrastive Learning

Lingyu Du*, Xucong Zhang, Guohao Lan

*Corresponding author for this work

Research output: Contribution to journalArticleScientificpeer-review

16 Downloads (Pure)

Abstract

Gaze estimation is of great importance to many scientific fields and daily applications, ranging from fundamental research in cognitive psychology to attention-aware systems. While recent advancements in deep learning have led to highly accurate gaze estimation systems, these solutions often come with high computational costs and depend on large-scale labeled gaze data for supervised learning, posing significant practical challenges. To move beyond these limitations, we present EfficientGaze, a resource-efficient framework for gaze representation learning. We introduce the frequency-domain gaze estimation, which exploits the feature extraction capability and the spectral compaction property of discrete cosine transform to substantially reduce the computational cost of gaze estimation systems for both calibration and inference. Moreover, to overcome the data labeling hurdle, we design a novel multi-task gaze-aware contrastive learning framework to learn gaze representations that are generic across subjects in an unsupervised manner. Our evaluation on two gaze estimation datasets demonstrates that EfficientGaze achieves comparable gaze estimation performance to existing supervised learning-based approaches, while enabling up to 6.80 times and 1.67 times speedup in system calibration and gaze estimation, respectively.

Original languageEnglish
Article number36
Number of pages24
JournalACM Transactions on Sensor Networks
Volume21
Issue number4
DOIs
Publication statusPublished - 2025

Keywords

  • Contrastive learning
  • gaze estimation
  • resource-efficient learning
  • unsupervised learning

Fingerprint

Dive into the research topics of 'Resource-efficient Gaze Estimation via Frequency-domain Multi-task Contrastive Learning'. Together they form a unique fingerprint.

Cite this