Multi-modal and Cross-Modal for Lecture Videos Retrieval - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2014

Multi-modal and Cross-Modal for Lecture Videos Retrieval

Résumé

The problem of multi-modal and cross-modal lecture videos retrieval is studied in this paper, on the basis of the use of document analysis techniques. In the context of this paper, a lecture video is represented by a set of subjects, in which a subject is represented by a Bag of mixed words -visual words and textual words-, each of them coming from speech recognition and OCR engines. Our work relies on two assumptions 1) a video may contain multiple subjects, 2) multiple modalities exist in the same lecture video document. We propose in this research a combination of technologies issuing from image document analysis and text mining. Visual words and textual words in images of lecture slides are extracted based on text detection and graphics localization computed on the sequences captured with a camera. Assuming that a subject in the video composes of a set of slides, lecture slides are clustered in different groups representing different possible subjects by using mixed words extracted. Multimodal and cross-modal lecture video retrieval are realized by the Bag of Subjects model. We discuss the proposed indexing and retrieval approach for lecture videos and report a quantitative evaluation on lecture videos of our University. It is shown that using Bag of Subjects for lecture video retrieval improves the retrieval accuracy.
Fichier non déposé

Dates et versions

hal-01247962 , version 1 (23-12-2015)

Identifiants

Citer

Nhu-Van Nguyen, Mickaël Coustaty, Jean-Marc Ogier. Multi-modal and Cross-Modal for Lecture Videos Retrieval. Pattern Recognition (ICPR), 2014 22nd International Conference on , Aug 2014, Stockholm, Sweden. pp.2667 - 2672, ⟨10.1109/ICPR.2014.461⟩. ⟨hal-01247962⟩

Collections

L3I UNIV-ROCHELLE
32 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More