Modeling Perception-Action Loops: Comparing Sequential Models with Frame-Based Classifiers - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2014

Modeling Perception-Action Loops: Comparing Sequential Models with Frame-Based Classifiers

Résumé

Modeling multimodal perception-action loops in face-to-face interactions is a crucial step in the process of building sensory-motor behaviors for social robots or users-aware Embodied Conversational Agents (ECA). In this paper, we compare trainable behavioral models based on sequential models (HMMs) and classifiers (SVMs and Decision Trees) inherently inappropriate to model sequential aspects. These models aim at giving pertinent perception/action skills for robots in order to generate optimal actions given the perceived actions of others and joint goals. We applied these models to parallel speech and gaze data collected from interacting dyads. The challenge was to predict the gaze of one subject given the gaze of the interlocutor and the voice activity of both. We show that Incremental Discrete HMM (IDHMM) generally outperforms classifiers and that injecting input context in the modeling process significantly improves the performances of all algorithms.
Fichier principal
Vignette du fichier
HAI_2014.pdf (817.69 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01061454 , version 1 (05-09-2014)

Identifiants

  • HAL Id : hal-01061454 , version 1

Citer

Alaeddine Mihoub, Gérard Bailly, Christian Wolf. Modeling Perception-Action Loops: Comparing Sequential Models with Frame-Based Classifiers. HAI 2014 - 2nd International Conference on Human-Agent Interaction, Oct 2014, Tsukuba, Japan. pp.309-314. ⟨hal-01061454⟩
475 Consultations
200 Téléchargements

Partager

Gmail Facebook X LinkedIn More