A. Moubayed, S. Edlund, J. Beskow, and J. , Taming Mona Lisa: communicating gaze faithfully in 2D and 3D facial projections, ACM Transactions on Interactive Intelligent Systems, vol.1, issue.2, 2012.

A. Moubayed, S. Skantze, G. Beskow, and J. , Lip-reading: Furhat audiovisual intelligibility of a back-projected animated face. Intelligent Virtual Agents-Lecture Notes in Computer Science, vol.7502, pp.196-203, 2012.

I. Albrecht, J. Haber, and H. Seidel, Automatic Generation of Non-Verbal Facial Expressions from Speech, Advances in Modelling, Animation and Rendering, pp.283-293, 2002.

P. D. Allopenna, J. S. Magnuson, and M. K. Tanenhaus, Tracking the time course of spoken word recognition using eye movements: Evidence for continuous mapping models, Journal of Memory and Language, vol.38, issue.4, pp.419-439, 1998.

F. Alnajar, T. Gevers, R. Valenti, and S. Ghebreab, Calibration-free gaze estimation using human gaze patterns, Presented at the Computer Vision (ICCV), 2013 IEEE International Conference on, pp.137-144, 2013.

G. Bailly, F. Elisei, S. Raidt, A. Casari, and A. Picot, Embodied conversational agents : computing and rendering realistic gaze patterns, Pacific Rim Conference on Multimedia Processing, vol.4261, pp.9-18, 2006.
URL : https://hal.archives-ouvertes.fr/hal-00143624

G. Bailly, F. Elisei, and M. Sauze, Beaming the gaze of a humanoid robot, HumanRobot Interaction (HRI), pp.47-48, 2015.
URL : https://hal.archives-ouvertes.fr/hal-01110288

G. Bailly, S. Raidt, and F. Elisei, Gaze, conversational agents and face-to-face communication, Speech Communication-Special Issue on Speech and Face-to-Face Communication, vol.52, issue.3, pp.598-612, 2010.
URL : https://hal.archives-ouvertes.fr/hal-00480335

R. Bajcsy, Active Perception. IEEE, Special Issue on Computer Vision, vol.76, pp.996-1005, 1988.

I. Barisic, B. Timmermans, U. Pfeiffer, G. Bente, K. Vogeley et al., In it together: using dual eyetracking to investigate real-time social interactions, Presented at the Proceedings from SIGCHI Conference on Human Factors in Computing Systems, 2013.

S. Baron-cohen, T. Jollife, C. Mortimore, and M. Robertson, Another advanced test of theory of mind: evidence from very high functioning adults with autism or Asperger syndrome, Journal of Child Psychology and Psychiatry, vol.38, issue.7, pp.813-822, 1997.

Y. Bengio and P. Frasconi, Input-output HMMs for sequence processing, IEEE Transactions on Neural Networks, vol.7, issue.5, pp.1231-1249, 1996.

,

C. Benoît, M. Grice, and V. Hazan, The SUS test: A method for the assessment of text-to-speech synthesis intelligibility using Semantically Unpredictable Sentences, Speech Communication, vol.18, pp.381-392, 1996.

M. Bindemann, A. M. Burton, I. C. Hooge, R. Jenkins, and E. F. De-haan, Faces retain attention, Psychonomic Bulletin & Review, vol.12, issue.6, pp.1048-1053, 2005.

S. M. Boker, J. F. Cohn, B. Theobald, I. Matthews, T. R. Brick et al., Effects of damping head movement and facial expression in dyadic conversation using real-time facial expression tracking and synthesized avatars, Philosophical Transactions of the Royal Society-Biological Sciences, vol.364, pp.3485-3495, 1535.

A. Borji, D. N. Sihite, and L. Itti, Quantitative Analysis of Human-Model Agreement in Visual Saliency Modeling: A Comparative Study. Image Processing, IEEE Transactions on, vol.22, issue.1, pp.55-69, 2013.

G. Brône and B. Oben, InSight Interaction: a multimodal and multifocal dialogue corpus. Language Resources and Evaluation, vol.49, pp.195-214, 2015.

,

J. N. Buchan, M. Paré, and K. G. Munhall, Spatial statistics of gaze fixations during dynamic face processing, Social Neuroscience, vol.2, issue.1, pp.1-13, 2007.

J. Carletta, R. L. Hill, C. Nicol, T. Taylor, J. P. De-ruiter et al., Eyetracking for two-person tasks with manipulation of a virtual world, Behavior Research Methods, vol.42, issue.1, pp.254-265, 2010.

H. H. Clark, Pointing and placing. Pointing: Where Language, Culture, and Cognition Meet, pp.243-268, 2003.

G. F. Cooper and E. Herskovits, A Bayesian method for the induction of probabilistic networks from data, Machine Learning, vol.9, pp.309-347, 1992.

,

A. Coutrot and N. Guyader, How saliency, faces, and sound influence gaze in dynamic social scenes, Journal of Vision, vol.14, issue.8, p.5, 2014.
URL : https://hal.archives-ouvertes.fr/hal-01018237

A. Coutrot, N. Guyader, G. Ionescu, and A. Caplier, Influence of soundtrack on eye movements during video exploration, Journal of Eye Movement Research, vol.5, issue.4, 2012.
URL : https://hal.archives-ouvertes.fr/hal-00723883

R. H. Cuijpers and D. Van-der-pol, Region of eye contact of humanoid Nao robot is similar to that of a human, Social Robotics, vol.8239, pp.280-289, 2013.

F. Cummins, Gaze and blinking in dyadic conversation: A study in coordinated behaviour among individuals. Language and Cognitive Processes, vol.27, pp.1525-1549, 2012.

R. Dale, R. Fusaroli, N. Duran, and D. C. Richardson, The self-organization of human interaction, Psychology of Learning and Motivation, vol.59, pp.43-95, 2013.

I. De-kok, Listening heads, 2013.

F. Delaunay, J. Greeff, and T. Belpaeme, A study of a retro-projected robotic face and its effectiveness for gaze reading by humans, ACM/IEEE International Conference on Human-Robot Interaction (HRI), pp.39-44, 2010.

R. Donat, L. Bouillaut, P. Aknin, and P. Leray, Reliability analysis using graphical duration models, Presented at the Availability, Reliability and Security, pp.795-800, 2008.
URL : https://hal.archives-ouvertes.fr/hal-00412299

S. Duffner and C. Garcia, Visual Focus of Attention estimation with unsupervised incremental learning, IEEE Transactions on Circuits and Systems for Video Technology, 2015.
URL : https://hal.archives-ouvertes.fr/hal-01153969

F. Elisei, G. Bailly, and A. Casari, Towards eyegaze-aware analysis and synthesis of audiovisual speech, Auditory-visual Speech Processing, pp.120-125, 2007.
URL : https://hal.archives-ouvertes.fr/hal-00169556

T. Hilvarenbeek and . Netherlands,

J. F. Ferreira, J. Lobo, P. Bessiere, M. Castelo-branco, and J. Dias, A Bayesian framework for active artificial perception, IEEE Transactions on Cybernetics, vol.43, issue.2, pp.699-711, 2013.
URL : https://hal.archives-ouvertes.fr/hal-00747148

F. Foerster, G. Bailly, and F. Elisei, Impact of iris size and eyelids coupling on the estimation of the gaze direction of a robotic talking head by human viewers, 2015.
URL : https://hal.archives-ouvertes.fr/hal-01228887

T. Foulsham, E. Walker, and A. Kingstone, The where, what and when of gaze allocation in the lab and the natural environment, Vision Research, vol.51, issue.17, pp.1920-1931, 2011.

J. H. Fuller, Head movement propensity, Experimental Brain Research, vol.92, issue.1, pp.152-164, 1992.

F. Mora, K. A. Odobez, and J. , Geometric generative gaze estimation (G3E) for remote RGB-D cameras, Presented at the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.1773-1780, 2014.

R. Fusaroli and K. Tylén, Investigating conversational dynamics: Interactive alignment, Interpersonal synergy, and collective task performance, Cognitive Science, vol.40, issue.1, pp.145-171, 2016.

M. Garau, M. Slater, S. Bee, and M. A. Sasse, The impact of eye gaze on communication using humanoid avatars, SIGCHI conference on Human factors in computing systems, pp.309-316, 2001.

S. Goferman, L. Zelnik-manor, and A. Tal, Context-aware saliency detection. Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol.34, issue.10, pp.1915-1926, 2012.

R. Gregory, Eye and Brain: The Psychology of Seeing, 1997.

E. Gu and N. I. Badler, Visual attention and eye gaze during multiparty conversations with distractions, Presented at the Intelligent Virtual Agents, pp.193-204, 2006.

D. A. Hanes and G. Mccollum, Variables contributing to the coordination of rapid eye/head gaze shifts, Biological Cybernetics, vol.94, pp.300-324, 2006.

J. M. Henderson, G. L. Malcolm, and C. Schandl, Searching in the dark: Cognitive relevance drives attention in real-world scenes, Psychonomic Bulletin & Review, vol.16, issue.5, pp.850-856, 2009.

J. K. Hietanen, Does your gaze direction and head orientation shift my visual attention?, Neuroreport, vol.10, issue.16, pp.3443-3447, 1999.

S. Hochreiter and J. Schmidhuber, Long short-term memory, Neural Computation, vol.9, issue.8, pp.1735-1780, 1997.

C. Huang and B. Mutlu, Learning-based Modeling of Multimodal Behaviors for Humanlike Robots, Proceedings of the 2014 ACM/IEEE International Conference on Human-robot Interaction, pp.57-64, 2014.

R. Ishii, K. Otsuka, S. Kumano, and J. Yamato, Analysis and modeling of next speaking start timing based on gaze behavior in multi-party meetings, IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.694-698, 2014.

L. Itti, N. Dhavale, and F. Pighin, Realistic avatar eye and head animation using a neurobiological model of visual attention, SPIE 48th Annual International Symposium on Optical Science and Technology, vol.5200, pp.64-78, 2003.

L. Itti, N. Dhavale, and F. Pighin, Photorealistic attention-based gaze animation, IEEE International Conference on Multimedia and Expo, pp.521-524, 2006.

F. Jensen, S. Lauritzen, and K. Olesen, Bayesian updating in recursive graphical models by local computations, Computational Statistics Quaterly, vol.4, issue.1, pp.269-282, 1990.

A. Karpathy, G. Toderici, S. Shetty, T. Leung, R. Sukthankar et al., Largescale video classification with convolutional neural networks, Presented at the Computer Vision and Pattern Recognition (CVPR), 2014 IEEE Conference on, pp.1725-1732, 2014.

H. Kobayashi and S. Kohshima, Unique morphology of the human eye and its adaptive meaning: comparative studies on external morphology of the primate eye, Journal of Human Evolution, vol.40, issue.5, pp.419-435, 2001.

D. Koller and N. Friedman, Probabilistic Graphical Models: Principles and Techniques-Adaptive Computation and Machine Learning, 2009.

A. Krizhevsky, I. Sutskever, and G. E. Hinton, ImageNet Classification with Deep Convolutional Neural Networks, Advances in Neural Information Processing (NIPS), 2012.

K. E. Laidlaw, T. Foulsham, G. Kuhn, and A. Kingstone, Social attention to a live person is critically different than looking at a videotaped person, Proc. Natl. Acad. Sci. (PNAS), vol.108, pp.5548-5553, 2011.

J. Lakin, V. Jefferis, C. Cheng, and T. Chartrand, The chameleon effect as social glue: evidence for the evolutionary significance of nonconscious mimicry, Nonverbal Behavior, vol.27, issue.3, pp.145-162, 2003.

S. R. Langton, The mutual influence of gaze and head orientation in the analysis of social attention direction, Quarterly Journal of Experimental Psychology, vol.53, issue.3, pp.825-845, 2000.

S. R. Langton, H. Honeyman, and E. Tessler, The influence of head contour and nose angle on the perception of eye-gaze direction, Perception & Psychophysics, vol.66, issue.5, pp.752-771, 2004.

C. R. Lansing and G. W. Mcconkie, Attention to facial regions in segmental and prosodic visual speech perception tasks, Journal of Speech, Language, and Hearing Research, vol.42, issue.3, pp.526-539, 1999.

S. P. Lee, J. B. Badler, and N. Badler, Eyes alive, ACM Transaction on Graphics, vol.21, issue.3, pp.637-644, 2002.

V. Levenshtein, Binary Codes Capable of Correcting Deletions, Insertions and Reversals. Soviet Physics Doklady, vol.10, issue.8, pp.707-710, 1966.

J. Li, Y. Tian, and T. Huang, Visual saliency with statistical priors, International Journal of Computer Vision, vol.107, issue.3, pp.239-253, 2014.

S. Liang, S. Fuhrman, and R. Somogyi, Reveal, a general reverse engineering algorithm for inference of genetic network architectures, Pacific symposium on biocomputing, vol.3, pp.18-29, 1998.

L. Marschner, S. Pannasch, J. Schulz, and S. Graupner, Social communication with virtual agents: The effects of body and gaze direction on attention and emotional responding in human observers, International Journal of Psychophysiology, vol.97, issue.2, pp.85-92, 2015.

D. Mcneill, Hand and Mind. What Gestures Reveal about Thought, 1992.

A. Mihoub, G. Bailly, and C. Wolf, Modelling perception-action loops: comparing sequential models with frame-based classifiers, Human-Agent Interaction (HAI), pp.309-314, 2014.

A. Mihoub, G. Bailly, and C. Wolf, Learning multimodal behavioral models for faceto-face social interaction, Journal on Multimodal User Interfaces (JMUI), vol.9, issue.3, pp.195-210, 2015.

A. Mihoub, G. Bailly, C. Wolf, and F. Elisei, Learning multimodal behavioral models for face-to-face social interaction, Journal on Multimodal User Interfaces, pp.1-16, 2015.
URL : https://hal.archives-ouvertes.fr/hal-01170991

A. Mihoub, G. Bailly, C. Wolf, and F. Elisei, Graphical models for social behavior modeling in face-to face interaction, Pattern Recognition Letters, vol.74, pp.82-89, 2016.
URL : https://hal.archives-ouvertes.fr/hal-01279427

K. Murphy, Dynamic bayesian networks: representation, inference and learning, UC Berkeley, 2002.

K. P. Murphy, The Bayes Net Toolbox for MATLAB, Computing Science and Statistics, vol.33, 2001.

B. Mutlu, T. Kanda, J. Forlizzi, J. Hodgins, and H. Ishiguro, Conversational gaze mechanisms for humanlike robots, ACM Transactions on Interactive Intelligent Systems (TiiS), vol.1, issue.2, p.12, 2012.

N. Neverova, C. Wolf, G. W. Taylor, and F. Nebout, ModDrop: adaptive multi-modal gesture recognition, IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), vol.38, pp.1692-1706, 2016.
URL : https://hal.archives-ouvertes.fr/hal-01178733

J. Ngiam, A. Khosla, M. Kim, J. Nam, H. Lee et al., Presented at the International conference on machine learning (ICML), pp.689-696, 2011.

D. Nguyen, . Bailly, . Gérard, . Elisei, and . Frédéric, Conducting neuropsychological tests with a humanoid robot: design and evaluation, IEEE International Conference on Cognitive Infocommunications-CogInfoCom. Wroclaw, 2016.
URL : https://hal.archives-ouvertes.fr/hal-01385666

T. Onuki, T. Ishinoda, Y. Kobayashi, and Y. Kuno, Designing robot eyes for gaze communication, IEEE Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV), pp.97-102, 2013.

K. Otsuka, Multimodal Conversation Scene Analysis for Understanding People's Communicative Behaviors in Face-to-Face Meetings, International Conference on Human-Computer Interaction (HCI), vol.12, pp.171-179, 2011.

K. Otsuka, Y. Takemae, and J. Yamato, A probabilistic inference of multipartyconversation structure based on Markov-switching models of gaze patterns, head directions, and utterances, International Conference on Multimodal Interfaces (ICMI), pp.191-198, 2005.

O. Oyekoya, A. Steed, and W. Steptoe, Eyelid kinematics for virtual characters, Computer Animation and Virtual Worlds, vol.21, issue.3-4, pp.161-171, 2010.

C. Pelachaud and M. Bilvi, Modelling gaze behavior for conversational agents, International Working Conference on Intelligent Virtual Agents, vol.2792, 2003.

A. S. Pentland, Social dynamics: Signals and behavior, Presented at the International Conference on Developmental Learning, 2004.

A. S. Pentland, Social Signal Processing, IEEE Signal Processing Magazine, vol.24, issue.4, pp.108-111, 2007.

A. Picot, G. Bailly, F. Elisei, and S. Raidt, Scrutinizing natural scenes: controlling the gaze of an embodied conversational agent, International Conference on Intelligent Virtual Agents (IVA), pp.272-282, 2007.
URL : https://hal.archives-ouvertes.fr/hal-00170337

S. Raidt, G. Bailly, and F. Elisei, Mutual gaze during face-to-face interaction, Auditory-visual Speech Processing, 2007.
URL : https://hal.archives-ouvertes.fr/hal-00169566

D. C. Richardson, R. Dale, and N. Z. Kirkham, The art of conversation is coordination common ground and the coupling of eye movements during dialogue, Psychological Science, vol.18, issue.5, pp.407-413, 2007.

D. C. Richardson, R. Dale, and K. Shockley, Synchrony and swing in conversation: coordination, temporal dynamics, and communication, Embodied Communication, pp.75-93, 2008.

E. F. Risko, K. E. Laidlaw, M. Freeth, T. Foulsham, and A. Kingstone, Social attention with real versus reel stimuli: toward an empirical approach to concerns about ecological validity, Frontiers in Human Neuroscience, vol.6, p.143, 2012.

E. F. Risko, D. C. Richardson, and A. Kingstone, Breaking the Fourth Wall of Cognitive Science Real-World Social Attention and the Dual Function of Gaze, Current Directions in Psychological Science, vol.25, issue.1, pp.70-74, 2016.

K. Ruhland, S. Andrist, J. Badler, C. Peters, N. Badler et al., Look me in the eyes: A survey of eye and gaze animation for virtual agents and artificial systems, pp.69-91, 2014.
URL : https://hal.archives-ouvertes.fr/hal-01025241

H. Sak, O. Vinyals, G. Heigold, A. Senior, E. Mcdermott et al., Sequence discriminative distributed training of long short-term memory recurrent neural networks, Entropy, vol.15, issue.16, pp.17-18, 2014.

B. Schauerte and R. Stiefelhagen, Look at this!" learning to guide visual saliency in human-robot interaction, 2014 IEEE/RSJ International Conference on, pp.995-1002, 2014.

R. Schmidt, S. Morr, P. Fitzpatrick, and M. J. Richardson, Measuring the dynamics of interactional synchrony, Journal of Nonverbal Behavior, vol.36, issue.4, pp.263-279, 2012.

A. Senju and T. Hasegawa, Direct gaze captures visuospatial attention, Vision Cognition, vol.12, pp.127-144, 2005.

Y. Sugano, Y. Matsushita, and Y. Sato, Appearance-based gaze estimation using visual saliency. Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol.35, issue.2, pp.329-341, 2013.

Y. Sun, Hierarchical object-based visual attention for machine vision (Thesis), 2003.

C. Teufel, D. M. Alexis, N. S. Clayton, and G. Davis, Mental-state attribution drives rapid, reflexive gaze following, Perception, & Psychophysics, vol.72, issue.3, pp.695-705, 2010.

M. Tomasello, Origins of Human Communication, 2008.

M. Tomasello, B. Hare, H. Lehmann, and J. Call, Reliance on head versus eyes in the gaze following of great apes and human infants: the cooperative eye hypothesis, Journal of Human Evolution, vol.52, pp.314-320, 2007.

G. Trabelsi, P. Leray, M. Ben-ayed, and A. M. Alimi, Benchmarking dynamic Bayesian network structure learning algorithms, Presented at the Modeling, Simulation and Applied Optimization (ICMSAO), 2013 5th International Conference on, pp.1-6, 2013.
URL : https://hal.archives-ouvertes.fr/hal-00812171

L. C. Trutoiu, E. J. Carter, I. Matthews, and J. K. Hodgins, Modeling and animating eye blinks, ACM Transactions on Applied Perception (TAP), vol.8, issue.3, pp.1-17, 2011.

R. Valenti and T. Gevers, Accurate eye center location through invariant isocentric patterns. Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol.34, issue.9, pp.1785-1798, 2012.

E. Van-der-burg, C. N. Olivers, A. W. Bronkhorst, and J. Theeuwes, Poke and pop: Tactile-visual synchrony increases visual saliency, Neuroscience Letters, vol.450, issue.1, pp.60-64, 2009.

E. Vatikiotis-bateson, I. Eigsti, S. Yano, and K. G. Munhall, Eye movement of perceivers during audiovisual speech perception, Perception & Psychophysics, vol.60, pp.926-940, 1998.

R. Vertegaal, R. Slagter, G. Van-der-veer, and A. Nijholt, Eye gaze patterns in conversations: There is more to conversational agents than meets the eyes, Conference on Human Factors in Computing Systems, pp.301-308, 2001.

V. Vinayagamoorthy, M. Garau, A. Steed, and M. Slater, An eye gaze model for dyadic interaction in an immersive virtual environment: Practice and experience, The Computer Graphics Forum, vol.23, issue.1, pp.1-11, 2004.

M. L. Võ, T. J. Smith, P. K. Mital, and J. M. Henderson, Do the eyes really have it? Dynamic allocation of attention when viewing moving faces, Journal of Vision, vol.13, issue.3, pp.1-14, 2012.

A. L. Yarbus, Eye movements during perception of complex objects, Eye Movements and Vision, vol.VII, pp.171-196, 1967.