Conférence
Notice
Lieu de réalisation
Journée GdR IASIS "Attention visuelle : prédiction et applications", INSA Rennes
Langue :
Anglais
Crédits
Jean-Marc Odobez (Intervention)
Conditions d'utilisation
Droit commun de la propriété intellectuelle
DOI : 10.60527/80mz-8771
Citer cette ressource :
Jean-Marc Odobez. GdR IASIS. (2024, 23 mai). Exploring Visual Attention: Methods for Analyzing Gaze cues in Everyday Contexts. [Vidéo]. Canal-U. https://doi.org/10.60527/80mz-8771. (Consultée le 16 mai 2025)

Exploring Visual Attention: Methods for Analyzing Gaze cues in Everyday Contexts

Réalisation : 23 mai 2024 - Mise en ligne : 8 octobre 2024
  • document 1 document 2 document 3
  • niveau 1 niveau 2 niveau 3
Descriptif

Beyond words, non-verbal behaviors (NVB) are known to play important roles in face-to-face interactions. However, decoding NVB is a challenging problem that involves both extracting subtle physical NVB cues and mapping them to higher-level communication behaviors or social constructs. Gaze, in particular, serves as a fundamental indicator of attention and interest, influencing communication and social signaling across various domains such as human-computer interaction, robotics, and medical diagnosis, notably in Autism Spectrum Disorders (ASD) assessment.

However, estimating others' visual attention, encompassing their gaze and Visual Focus of Attention (VFOA), remains highly challenging, even for humans. It requires not only inferring accurate 3D gaze directions but also understanding the contextual scene to discern which object in the field of view is actually looked at. Context can include people activities that can provide priors about which objects are looked at, or the scene structure to detect obstructions in the line of sight. Recent research has pursued two avenues to address this: the first one focused on improving appearance-based 3D gaze estimation from images and videos, while the second investigated gaze following ?the task of inferring where a person looks in an image.

This presentation will explore ideas and methodologies addressing both challenges. Initially, it delves into advancements in 3D gaze estimation, including personalized model construction via few-shot learning and gaze redirection eye synthesis, differential gaze estimation, and leveraging social interaction priors for model adaptation. Subsequently, recent models for estimating gaze targets in real-world settings are introduced, including the inference of social labels like eye contact and shared attention.

Intervention