Learning to detect dysarthria from raw speech

Abstract : Speech classifiers of paralinguistic traits traditionally learn from diverse hand-crafted low-level features, by selecting the relevant information for the task at hand. We explore an alternative to this selection, by learning jointly the classifier, and the feature extraction. Recent work on speech recognition has shown improved performance over speech features by learning from the waveform. We extend this approach to paralinguistic classification and propose a neural network that can learn a filterbank, a normalization factor and a compression power from the raw speech, jointly with the rest of the architecture. We apply this model to dysarthria detection from sentence-level audio recordings. Starting from a strong attention-based baseline on which mel-filterbanks outperform standard low-level descriptors, we show that learning the filters or the normalization and compression improves over fixed features by 10% absolute accuracy. We also observe a gain over OpenSmile features by learning jointly the feature extraction, the normalization, and the compression factor with the architecture. This constitutes a first attempt at learning jointly all these operations from raw audio for a speech classification task.
Complete list of metadatas

https://hal.archives-ouvertes.fr/hal-02274504
Contributor : Ewan Dunbar <>
Submitted on : Thursday, August 29, 2019 - 9:22:39 PM
Last modification on : Saturday, August 31, 2019 - 1:11:42 AM

Links full text

Identifiers

  • HAL Id : hal-02274504, version 1
  • ARXIV : 1811.11101

Collections

Citation

Juliette Millet, Neil Zeghidour. Learning to detect dysarthria from raw speech. IEEE International Conference on Acoustics, Speech and Signal Processing, May 2019, Brighton, United Kingdom. ⟨hal-02274504⟩

Share

Metrics

Record views

20