R. Francis, The game is a roleplaying game in the town of colonial Williamsburg during the American Revolution, Proc. of the American educational research association conference, vol.8, 2006.

C. Huang and Y. Huang, Annales school-based serious game creation framework for Taiwan indigenous cultural heritage, Journal of Computing in Cultural Heritage, vol.6, issue.2, 2013.

F. Cameron and S. Kenderdine, Theorizing Digital Cultural Heritage: A Critical Dis-course, 2010.

M. Ioannides, D. Fellner, A. Georgopoulos, and D. Hadjimitsis, Digital Heritage, Third International Conference, 2010.

N. Aikawa, An Historical Overview of the Preparation of the UNESCO International Convention for the Safeguarding of Intangible Heritage, Museum International, vol.56, pp.137-149, 2004.

V. Hafstein, Intangible Heritage as List: From Masterpieces to Representation, Intangible Heritage, pp.93-111, 2009.

P. Nas, Masterpieces of Oral and Intangible Heritage: Reflections on the UNESCO World Heritage List, Current Anthropology, vol.43, issue.1, pp.139-143, 2002.

M. Alivizatou, The UNESCO Programme for the Proclamation of Masterpieces of the Oral and Intangible Heritage of Humanity: A Critical Examination, Journal of Museum Ethnography, vol.19, pp.34-42, 2007.

L. Bolton, Unfolding the Moon: Enacting Women's Kastom in Vanuatu, 2003.

K. Huffman, The Fieldworkers of the Vanuatu Cultural Centre and their Contribution to the Audiovisual Collections in Bonnemaison, pp.290-293, 1996.

S. Zafeiriou and L. Yin, 3D facial behaviour analysis and understanding, Image and Vision Computing, vol.30, pp.681-682, 2012.

P. Ekman, R. Levenson, and W. Friesen, Emotions Differ in Autonomic Nervous System Activity, Science, vol.221, pp.1208-1210, 1983.

O. Engwall, Modeling of the vocal tract in three dimensions, Proceedings, Eu-rospeech99, p.113116, 1999.

S. Fels, J. E. Lloyd, K. Van-den-doel, F. Vogt, I. Stavness et al., Developing physically-based, dynamic vocal tract models using Artisynth, Proceedings of ISSP 6, vol.19, p.309320, 1991.

P. Badin, G. Bailly, L. Reveret, M. Baciu, C. Segebarth et al., Three-dimensional linear articulatory modeling of tongue, lips and face, based on MRI and video images, Journal of Phonetics, vol.30, issue.3, pp.533-553, 2002.
URL : https://hal.archives-ouvertes.fr/hal-00798689

M. Stone, A three-dimensional model of tongue movement based on ultrasound and x-ray microbeam data, The Journal of the Acoustical Society of America, vol.87, p.2207, 1990.

O. Engwall, From real-time MRI to 3D tongue movements, Proceedings, 8th International Conference on Spoken Language Processing (ICSLP), vol.2, pp.1109-1112, 2004.

M. Stone and A. Lundberg, Three-dimensional tongue surface shapes of English con-sonants and vowels, Journal of the Acoustical Society of America, vol.99, issue.6, p.37283737, 1996.

N. Henrich, B. Lortat-jacob, M. Castellengo, L. Bailly, and X. Pelorson, Period-doubling occurences in singing: the "bassu" case in traditional Sardinian "A Tenore" singing, Proceedings of the International Conference on Voice Physiology and Biomechanics, 2006.
URL : https://hal.archives-ouvertes.fr/hal-00371458

N. Henrich, L. Bailly, X. Pelorson, and B. Lortat-jacob, Physiological and physical understanding of singing voice practices: the Sardinian Bassu case, AIRS Start-up meeting, 2009.

W. Cho, J. Hong, and H. Park, Real-Time Ultrasonographic Assessment of True Vocal Fold Length in Professional Singers, Journal of Voice, vol.26, issue.6, pp.1-6, 2012.

G. Troup, T. Griffiths, M. Schneider-kolsky, and T. Finlayson, Ultrasound Observa-tion of Vowel Tongue Shapes in Trained Singers, Proceedings of the 30 th Condensed Matter and Materials Meeting, 2006.

M. Caon, Context-Aware 3D Gesture Interaction Based on Multiple Kinects, Proceedings of the First International Conference on Ambient Computing, Applications, Services and Technologies, pp.7-12, 2011.

T. Coduys, C. Henry, and A. Cont, TOASTER and KROONDE: high-resolution and high-speed realtime sensor interfaces, Proceedings of the Conference on New Interfaces for Musical Expression, pp.205-206, 2004.

F. Bevilacqua, B. Zamborlin, A. Sypniewski, N. Schnell, F. Guedy et al., Gesture in embodied communication and human-computer interaction, pp.73-84, 2010.

M. Boucher, Virtual Dance and Motion-Capture, Contemporary Aesthetics, vol.9, 2011.

R. Aylward and J. A. Paradiso, Sensemble: A Wireless, Compact, Multi-User Sensor System for Interactive Dance, Proceedings of the International Conference on New Interfaces for Musical Expression (NIME06), pp.134-139, 2006.

D. Drobny, M. Weiss, and J. Borchers, Saltate!: a sensor-based system to support dance beginners, Extended abstracts on Human factors in Computing Systems, Proceedings of the CHI 09 International Conference, pp.3943-3948, 2009.

F. Bevilacqua, L. Naugle, and C. Dobrian, Music control from 3D motion capture of dance, 2001.

C. Dobrian and F. Bevilacqua, Gestural control of music: using the vicon 8 motion capture system, Proceedings of the Conference on New Interfaces for Musical Expression (NIME), pp.161-163, 2003.

M. Raptis, D. Kirovski, and H. Hoppe, Real-time classification of dance gestures from skeleton animation, Proceedings of ACM SIGGRAPH/Eurographics Symposium on Computer Animation, pp.147-156, 2011.

D. S. Alexiadis, P. Kelly, P. Daras, N. E. O'connor, T. Boubekeur et al., Evaluating a dancer's performance using kinect-based skeleton tracking, Proceedings of the 19th ACM international conference on Multimedia, pp.659-662, 2011.

S. Essid, D. S. Alexiadis, R. Tournemenne, M. Gowing, P. Kelly et al., An advanced virtual dance performance evaluator, Proceedings of the 37th International Conference on Acoustics, Speech, and Signal Processing, pp.2269-2272, 2012.

G. Alankus, A. A. Bayazit, and O. B. Bayazit, Automated motion synthesis for dancing characters: Motion Capture and Retrieval, Comput. Animat. Virtual Worlds, vol.16, issue.3-4, pp.259-271, 2005.

D. Bouchard and N. Badler, Semantic segmentation of motion capture using laban movement analysis,Intelligent Virtual Agents, pp.37-44, 2007.

K. Kahol, P. Tripathi, and S. Panchanathan, Automated gesture segmentation from dance sequences, Proceedings of the Sixth IEEE International Conference on Automatic Face and Gesture Recognition (FGR04), pp.883-888, 2004.

J. James, T. Ingalls, G. Qian, L. Olsen, D. Whiteley et al., Movement-based interactive dance performance, Proceedings of the 14th annual ACM International Conference on Multimedia, pp.470-480, 2006.

M. Malempre, Pour une poignee de danses, Dapo Hainaut, 2010.

T. Calvert, W. Wilke, R. Ryman, and I. Fox, Applications of computers to dance, Computer Graphics and Applications, vol.25, issue.2, pp.6-12, 2005.

Y. Shen, X. Wu, C. Lua, and H. Cheng, National Dances Protection Based on Motion Capture Technology, vol.51, pp.78-81, 2012.

W. M. Brown, L. Cronk, K. Grochow, A. Jacobson, C. K. Liu et al., Dance reveals symmetry especially in young men, Nature, vol.438, issue.7071, pp.1148-1150, 2005.

D. Tardieu, X. Siebert, B. Mazzarino, R. Chessini, J. Dubois et al., Browsing a dance video collection: dance analysis and interface design, Journal on Multimodal User Interfaces, vol.4, issue.1, pp.37-46, 2010.

J. C. Chan, H. Leung, J. K. Tang, and T. Komura, A virtual reality dance training system using motion capture technology, Learning Technologies, IEEE Transactions, vol.4, issue.2, pp.187-195, 2011.

I. Cohen, A. Garg, and T. Huang, Emotion Recognition from Facial Expression Using Multilevel HMM, Proceedings of the Neural Information Processing Systems Workshop on Affective Computing, 2000.

F. Bourel, C. Chibelushi, and A. Low, Robust Facial Expression Recognition Using a State-Based Model of Spatially-Localized Facial Dynamics, Proceedings of IEEE International Conference on Automatic Face and Gesture Recognition, 2002.

B. Schuller, S. Reiter, R. Mueller, A. Hames, and G. Rigoll, Speaker Independent Speech Emotion Recognition by Ensemble Classification, Proceedings of the IEEE International Conference on Multimedia and Expo, pp.864-867, 2005.

C. Busso, Z. Deng, S. Yildirim, M. Bulut, C. Lee et al., Analysis of Emotional Recognition Using Facial Expressions, Speech and Multimodal Information, Proceedings of the International Conference on Multimodal Interfaces, pp.205-211, 2004.

R. Picard, E. Vyzas, and J. Healey, Toward Machine Emotional Intelligence: Analysis of Affective Physiological State, IEEE Trans. Pattern Analysis and Machine Intelligence, vol.23, issue.10, pp.1175-1191, 2001.

F. Nasoz, C. Lisetti, K. Alvarez, and N. Finkelstein, Emotion Recognition from Physiological Signals for User Modeling of Affect, Proceedings of the International Conference on User Modeling, 2003.

C. Lisetti and N. F. , Using Non-invasive Wearable Computers to Recognize Human Emotions from Physiological Signals, EURASIP Journal on Applied Signal Processing, vol.11, pp.1672-1687, 2004.

D. Mcintosh, A. Reichmann-decker, P. Winkielman, and J. Wilbarger, When the Social Mirror Breaks: Deficits in Automatic, But Not Voluntary, Mimicry of Emotional Facial Expressions in Autism, Developmental Science, vol.9, pp.295-302, 2006.

F. Esposito, D. Malerba, G. Semeraro, O. Altamura, S. Ferilli et al., Machine learning methods for automatically processing historical documents: from paper acquisition to XML transformation, Proceedings of the First International Workshop on Document Image Analysis for Libraries (DIAL, 04), pp.328-335, 2004.

A. Mallik, S. Chaudhuri, and H. Ghosh, Nrityakosha: Preserving the Intangible Heritage of Indian Classical Dance, ACM Journal on Computing and Cultural Heritage, vol.4, issue.3, p.11, 2011.

M. Makridis and P. Daras, Automatic classification of archaeological pottery sherds, Journal on Computing and Cultural Heritage (JOCCH), vol.5, issue.4, p.15, 2012.

A. Karasik, A complete, automatic procedure for pottery documentation and analysis, Proceedings of the IEEE Computer Vision and Pattern Recognition Workshops (CVPRW), pp.29-34, 2010.

S. Vrochidis, C. Doulaverakis, A. Gounaris, E. Nidelkou, L. Makris et al., A Hybrid Ontology and Visual-based Retrieval Model for Cultural Heritage Multimedia Collections, International Journal of Metadata, Semantics and Ontologies, vol.3, issue.3, pp.167-182, 2008.

M. Liggins, D. L. Hall, and J. Llina, Handbook of Multisensor Data Fusion, Theory and Practice, 2008.

O. Punska, Bayesian approach to multisensor data fusion, 1999.

S. Nikolopoulos, C. Lakka, I. Kompatsiaris, C. Varytimidis, K. Rapantzikos et al., Compound document analysis by fusing evidence across media, Proceedings of the International Workshop on Content-Based Multimedia Indexing, pp.175-180, 2009.

S. Chang, D. Ellis, W. Jiang, K. Lee, A. Yanagawa et al., Largescale multimodal semantic concept detection for consumer video, Proceedings of the international workshop on Workshop on multimedia information retrieval (MIR '07), pp.255-264, 2007.

R. Huber-mörk, S. Zambanini, M. Zaharieva, and M. Kampel, Identification of ancient coins based on fusion of shape and local features, Machine Vision and Applications, vol.22, pp.983-994, 2011.

D. Datcu and L. J. Rothkrantz, Semantic Audio-Visual Data Fusion for Automatic Emotion Recognition, Euromedia, 2008.

M. Koolen and J. Kamps, Searching cultural heritage data: Does structure help expert searchers?', Proceedings of RIAO '10 Adaptivity, Personalization and Fusion of Heterogeneous Information, pp.152-155, 2010.

L. Bai, .. S. Lao, W. Zhang, G. J. Jones, and A. F. Smeaton, Video Semantic, Content Analysis Framework Based on Ontology Combined MPEG-7', Adaptive Multimedia Retrieval: Retrieval, User, and Semantics, pp.237-250, 2007.

S. Dasiopoulou, V. Mezaris, I. Kompatsiaris, V. K. Papastathis, and G. M. Strintzis, KnowledgeAssisted Semantic Video Object Detection, IEEE Transactions on Circuits and Systems for Video Technology', Special Issue on Analysis and Understanding for Video Adaptation, vol.15, pp.1210-1224, 2005.

J. Lien, T. Kanade, J. Cohn, and C. Li, Automated Facial Expression Recognition Based on FACS Action Units, Proceedings of the Third IEEE Conference on Automatic Face and Gesture Recognition, pp.390-395, 1998.

P. Mulholland, A. Wolff, T. Collins, and Z. Zdrahal, An event-based approach to describing and understanding museum narratives, Proceedings: Detection, Representation, and Exploitation of Events in the Semantic Web Workshop in conjunction with theInternational Semantic Web Conference, 2011.

I. Kollia, V. Tzouvaras, N. Drosopoulos, and G. Stamou, A Systemic Approach for Effective Semantic Access to Cultural Content' Semantic Web -Interoperability, Usability, Applicability, vol.3, issue.1, pp.65-83, 2012.

A. Gaitatzes, D. Christopoulos, and M. Roussou, Reviving the past: cultural heritage meets virtual reality, Proceedings of the 2001 conference on Virtual reality, archeology, and cultural heritage, pp.103-110, 2001.

M. Ott and F. Pozzi, Problem-based learning: An instructional model and its constructivist framework, Computers in Human Behavior, vol.27, issue.4, pp.31-38, 1995.

K. H. Veltman, Challenges for ICT/UCT Applications in Cultural Heritage, ICT and Heritage, 2005.

M. Mortara, C. E. Catalano, F. Bellotti, G. Fiucci, M. Houry-panchetti et al., Learning Cultural Heritage by serious games, Journal of Cultural Heritage, vol.15, issue.3, pp.318-325, 2014.
URL : https://hal.archives-ouvertes.fr/hal-01120560

E. F. Anderson, L. Mcloughlin, F. Liarokapis, C. Peters, P. Petridis et al., Proceedings of the The 10th International Symposium on Virtual Reality, Archaeology and Cultural Heritage VAST -State of the Art Reports, 2009.

M. Ott and F. Pozzi, ICT and Cultural Heritage Education: Which Added Value, 2008.

, Emerging Technologies and Information Systems for the Knowledge Society, Lecture Notes in Computer Science, vol.5288, pp.131-138

X. Rodet, Y. Potard, and J. Barriere, The CHANT project: from the synthesis of the singing voice to synthesis in general, Computer Music Journal, vol.8, issue.3, pp.15-31, 1984.

G. Berndtsson, The KTH rule system for singing synthesi, Computer Music Journal, vol.20, issue.1, p.7691, 1996.

P. Cook, Physical Models for Music Synthesis, and a Meta-Controller for Real Time Performance, Proceedings of the International Computer Music Conference and Festival, 1992.

P. Cook, Singing voice synthesis: History, current work, and future directions, Computer Music Journal, vol.20, issue.3, p.3846, 1996.

G. Bennett and X. Rodet, Synthesis of the singing voice, Current directions in computer music research, pp.19-44, 1989.

H. Kenmochi and H. Ohshita, Vocaloid-commercial singing synthesizer based on sample concatenation. Presented at Interspeech, pp.4009-40010, 2007.

A. Burns and M. M. Wanderley, Visual methods for the retrieval of guitarist fingering, Proceedings of the Conference on New interfaces for musical expression, Pompidou: IRCAM-Centre, pp.196-199, 2006.

, Vision par ordinateur pour la reconnaissance des gestes musicaux des doigts, Revue Francophone d'Informatique Musicale, 2013.

D. Grunberg, Gesture Recognition for Conducting Computer Music, 2009.

J. Verner, MIDI guitar synthesis yesterday, today and tomorrow, an overview of the whole fingerpicking thing, Recording Magazine, vol.8, issue.9, pp.52-57, 1995.

C. Traube, An interdisciplinary study of the timbre of the classical guitar, 2004.

Y. Takegawa, T. Terada, and S. Nishio, Design and Implementation of a Real-time Fingering Detection System for Piano Performances, Proceedings of the International Computer Music Conference, pp.67-74, 2006.

J. Macritchie, B. Buck, and N. Bailey, Visualising musical structure through performance gesture, Proceedings of the International Society for Music Information Retrieval Conference, pp.237-242, 2009.

K. Dimitropoulos, S. Manitsaris, F. Tsalakanidou, S. Nikolopoulos, B. Denby et al.,

O. Charisis, A. Deroo, I. Manitsaris, N. Kompatsiaris, and . Grammalidis, Capturing the Intangible: An Introduction to the i-Treasures Project, Proc. 9th International Conference on Computer Vision Theory and Applications (VISAPP2014), pp.5-8, 2014.

M. Brand and A. Hertzmann, Style machines, Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques (SIGGRAPH 2000), pp.183-192, 2000.

G. Bennett and X. Rodet, Synthesis of the singing voice, Current directions in computer music research, pp.19-44, 1989.

A. Kitsikidis, K. Dimitropoulos, S. Douka, and N. Grammalidis, Dance analysis using multiple kinect sensors, Computer Vision Theory and Applications (VISAPP), 2014 International Conference on, vol.2, pp.789-795, 2014.