CamemBERT: a Tasty French Language Model - Archive ouverte HAL Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2019

CamemBERT: a Tasty French Language Model

Résumé

Pretrained language models are now ubiquitous in Natural Language Processing. Despite their success, most available models have either been trained on English data or on the concatenation of data in multiple languages. This makes practical use of such models—in all languages except English—very limited. Aiming to address this issue for French, we release CamemBERT, a French version of the Bi-directional Encoders for Transformers (BERT). We measure the performance of CamemBERT compared to multilingual models in multiple downstream tasks, namely part-of-speech tagging, dependency parsing, named-entity recognition, and natural language inference. CamemBERT improves the state of the art for most of the tasks considered. We release the pretrained model for CamemBERT hoping to foster research and downstream applications for French NLP.

Dates et versions

hal-02445946 , version 1 (20-01-2020)

Identifiants

Citer

Louis Martin, Benjamin Muller, Pedro Javier Ortiz Suárez, Yoann Dupont, Laurent Romary, et al.. CamemBERT: a Tasty French Language Model. 2019. ⟨hal-02445946⟩
858 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More