Hierarchical Pre-training for Sequence Labelling in Spoken Dialog - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2020

Hierarchical Pre-training for Sequence Labelling in Spoken Dialog

Résumé

Sequence labelling tasks like Dialog Act and Emotion/Sentiment identification are a key component of spoken dialog systems. In this work, we propose a new approach to learn generic representations adapted to spoken dialog, which we evaluate on a new benchmark we call Sequence labellIng evaLuatIon benChmark fOr spoken laNguagE benchmark (SILICONE). SILICONE is model-agnostic and contains 10 different datasets of various sizes. We obtain our representations with a hierarchical encoder based on transformer architectures, for which we extend two well-known pre-training objectives. Pre-training is performed on OpenSubtitles: a large corpus of spoken dialog containing over 2.3 billion of tokens. We demonstrate how hierarchical encoders achieve competitive results with consistently fewer parameters compared to state-of-the-art models and we show their importance for both pre-training and fine-tuning.
Fichier principal
Vignette du fichier
2009.11152 (3.17 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03134851 , version 1 (04-01-2024)

Identifiants

Citer

Emile Chapuis, Pierre Colombo, Matteo Manica, Matthieu Labeau, Chloé Clavel. Hierarchical Pre-training for Sequence Labelling in Spoken Dialog. Findings of the Association for Computational Linguistics: EMNLP 2020, Nov 2020, Online, France. pp.2636-2648, ⟨10.18653/v1/2020.findings-emnlp.239⟩. ⟨hal-03134851⟩
41 Consultations
4 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More