Learning Word Embeddings: Unsupervised Methods for Fixed-size Representations of Variable-length Speech Segments - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2018

Learning Word Embeddings: Unsupervised Methods for Fixed-size Representations of Variable-length Speech Segments

Résumé

Fixed-length embeddings of words are very useful for a variety of tasks in speech and language processing. Here we systematically explore two methods of computing fixed-length embeddings for variable-length sequences. We evaluate their susceptibility to phonetic and speaker-specific variability on English, a high resource language and Xitsonga, a low resource language, using two evaluation metrics: ABX word discrimination and ROC-AUC on same-different phoneme n-grams. We show that a simple downsampling method supplemented with length information can outperform the variable-length input feature representation on both evaluations. Recurrent autoencoders, trained without supervision, can yield even better results at the expense of increased computational complexity.
Fichier principal
Vignette du fichier
Holzenberger_DKRD_2018_fixed_length_embeddings_for_words.Interspeech.pdf (505.44 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01888708 , version 1 (07-12-2018)

Identifiants

Citer

Nils Holzenberger, Mingxing Du, Julien Karadayi, Rachid Riad, Emmanuel Dupoux. Learning Word Embeddings: Unsupervised Methods for Fixed-size Representations of Variable-length Speech Segments. Interspeech 2018, Sep 2018, Hyderabad, India. ⟨10.21437/Interspeech.2018-2364⟩. ⟨hal-01888708⟩
355 Consultations
722 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More