Using the FASST source separation toolbox for noise robust speech recognition - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2011

Using the FASST source separation toolbox for noise robust speech recognition

Résumé

We describe our submission to the 2011 CHiME Speech Separation and Recognition Challenge. Our speech separation algorithm was built using the Flexible Audio Source Separation Toolbox (FASST) we developed recently. This toolbox is an implementation of a general flexible framework based on a library of structured source models that enable the incorporation of prior knowledge about a source separation problem via user-specifiable constraints. We show how to use FASST to develop an efficient speech separation algorithm for the CHiME dataset. We also describe the acoustic model training and adaptation strategies we used for this submission. Altogether, as compared to the baseline system, we obtain an improvement of keyword recognition accuracies in all conditions. The best improvement of about 40 % is achieved in the worst condition of -6 dB Signal-to-Noise-Ratio (SNR), where 18 % of this improvement is due to the speech separation. The improvement decreases when the SNR increases. These results indicate that audio source separation can be very helpful to improve speech recognition in noisy or multi-source environments.
Fichier principal
Vignette du fichier
CHiME_submission_v4.pdf (104.57 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

inria-00598734 , version 1 (07-06-2011)

Identifiants

  • HAL Id : inria-00598734 , version 1

Citer

Alexey Ozerov, Emmanuel Vincent. Using the FASST source separation toolbox for noise robust speech recognition. International Workshop on Machine Listening in Multisource Environments (CHiME 2011), Sep 2011, Florence, Italy. ⟨inria-00598734⟩
226 Consultations
278 Téléchargements

Partager

Gmail Facebook X LinkedIn More