What do you mean, BERT? Assessing BERT as a Distributional Semantics Model - Archive ouverte HAL Accéder directement au contenu
Article Dans Une Revue Proceedings of the Society for Computation in Linguistics Année : 2020

What do you mean, BERT? Assessing BERT as a Distributional Semantics Model

Résumé

Contextualized word embeddings, i.e. vector representations for words in context, are naturally seen as an extension of previous non-contextual distributional semantic models. In this work, we focus on BERT, a deep neural network that produces contextualized embeddings and has set the state-of-the-art in several semantic tasks, and study the semantic coherence of its embedding space. While showing a tendency towards coherence, BERT does not fully live up to the natural expectations for a semantic vector space. In particular, we find that the position of the sentence in which a word occurs, while having no meaning correlates , leaves a noticeable trace on the word embeddings and disturbs similarity relationships.
Fichier principal
Vignette du fichier
1911.05758.pdf (399.83 Ko) Télécharger le fichier
Loading...

Dates et versions

hal-02484933 , version 1 (25-02-2020)

Identifiants

Citer

Timothee Mickus, Mathieu Constant, Denis Paperno, Kees van Deemter. What do you mean, BERT? Assessing BERT as a Distributional Semantics Model. Proceedings of the Society for Computation in Linguistics, 2020, 3, ⟨10.7275/t778-ja71⟩. ⟨hal-02484933⟩
95 Consultations
207 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More