Improving Domain Generalization in Appearance-Based Gaze Estimation With Consistency Regularization

Gaze estimation, a method for understanding human behavior by analyzing where a person is looking, has significant applications in various fields including advertising, driving assistance, medical diagnostics, and human-computer interaction. Although appearance-based methods have shown promising per...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2023, Vol.11, p.137948-137956
Hauptverfasser: Back, Moon-Ki, Yoo, Cheol-Hwan, Yoo, Jang-Hee
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Gaze estimation, a method for understanding human behavior by analyzing where a person is looking, has significant applications in various fields including advertising, driving assistance, medical diagnostics, and human-computer interaction. Although appearance-based methods have shown promising performance in uncontrolled environments, they often perform poorly when applied to similar but different domains due to variances in image quality, gaze distribution, and illumination. To overcome this limitation, this study aims to improve the domain generalization of appearance-based gaze estimation models using deep learning techniques. We propose an end-to-end deep learning approach that facilitates domain-agnostic feature learning and introduce a novel loss function, spherical gaze distance (SGD), and a regularization method, gaze consistency regularization (GCR). Our experiments, conducted using three commonly used datasets for appearance-based gaze estimation: ETH-XGaze, MPIIGaze, and GazeCapture, demonstrate the effectiveness of SGD and GCR. The results show that the proposed approach outperforms all the state-of-the-art methods on the domain generalization task and significantly improves performance when SGD and GCR are combined. These findings have important implications for the field of gaze estimation, suggesting that the proposed method could enhance the robustness and generalizability of gaze estimation models.
ISSN:2169-3536
2169-3536
DOI:10.1109/ACCESS.2023.3340446