Skip to Main content Skip to Navigation
New interface
Conference papers

Learning to detect dysarthria from raw speech

Abstract : Speech classifiers of paralinguistic traits traditionally learn from diverse hand-crafted low-level features, by selecting the relevant information for the task at hand. We explore an alternative to this selection, by learning jointly the classifier, and the feature extraction. Recent work on speech recognition has shown improved performance over speech features by learning from the waveform. We extend this approach to paralinguistic classification and propose a neural network that can learn a filterbank, a normalization factor and a compression power from the raw speech, jointly with the rest of the architecture. We apply this model to dysarthria detection from sentence-level audio recordings. Starting from a strong attention-based baseline on which mel-filterbanks outperform standard low-level descriptors, we show that learning the filters or the normalization and compression improves over fixed features by 10% absolute accuracy. We also observe a gain over OpenSmile features by learning jointly the feature extraction, the normalization, and the compression factor with the architecture. This constitutes a first attempt at learning jointly all these operations from raw audio for a speech classification task.
Complete list of metadata
Contributor : Ewan Dunbar Connect in order to contact the contributor
Submitted on : Thursday, August 29, 2019 - 9:22:39 PM
Last modification on : Friday, November 18, 2022 - 9:23:27 AM

Links full text


  • HAL Id : hal-02274504, version 1
  • ARXIV : 1811.11101


Juliette Millet, Neil Zeghidour. Learning to detect dysarthria from raw speech. ICASSP-2019 - IEEE International Conference on Acoustics, Speech and Signal Processing, May 2019, Brighton, United Kingdom. ⟨hal-02274504⟩



Record views