Contextualized word embeddings have been employed effectively across several tasks in Natural Language Processing, as they have proved to carry useful semantic information. However, it is still hard to link them to structured sources of knowledge. In this paper we present ARES (context-AwaRe Embeddings of Senses), a semi-supervised approach to producing sense embeddings for the lexical meanings within a lexical knowledge base that lie in a space that is comparable to that of contextualized word vectors. ARES representations enable a simple 1 Nearest-Neighbour algorithm to outperform state-of-the-art models, not only in the English Word Sense Disambiguation task, but also in the multilingual one, whilst training on sense-annotated data in English only. We further assess the quality of our embeddings in the Word-in-Context task, where, when used as an external source of knowledge, they consistently improve the performance of a neural model, leading it to compete with other more complex architectures. ARES embeddings for all WordNet concepts and the automatically-extracted contexts used for creating the sense representations are freely available at http://sensembert.org/ares.

With More Contexts Comes Better Performance: Contextualized Sense Embeddings for All-Round Word Sense Disambiguation / Scarlini, Bianca; Pasini, Tommaso; Navigli, Roberto. - (2020), pp. 3528-3539. (Intervento presentato al convegno The 2020 Conference on Empirical Methods in Natural Language Processing tenutosi a Online) [10.18653/v1/2020.emnlp-main.285].

With More Contexts Comes Better Performance: Contextualized Sense Embeddings for All-Round Word Sense Disambiguation

Scarlini, Bianca
Primo
;
Pasini, Tommaso
Secondo
;
Navigli, Roberto
Ultimo
2020

Abstract

Contextualized word embeddings have been employed effectively across several tasks in Natural Language Processing, as they have proved to carry useful semantic information. However, it is still hard to link them to structured sources of knowledge. In this paper we present ARES (context-AwaRe Embeddings of Senses), a semi-supervised approach to producing sense embeddings for the lexical meanings within a lexical knowledge base that lie in a space that is comparable to that of contextualized word vectors. ARES representations enable a simple 1 Nearest-Neighbour algorithm to outperform state-of-the-art models, not only in the English Word Sense Disambiguation task, but also in the multilingual one, whilst training on sense-annotated data in English only. We further assess the quality of our embeddings in the Word-in-Context task, where, when used as an external source of knowledge, they consistently improve the performance of a neural model, leading it to compete with other more complex architectures. ARES embeddings for all WordNet concepts and the automatically-extracted contexts used for creating the sense representations are freely available at http://sensembert.org/ares.
2020
The 2020 Conference on Empirical Methods in Natural Language Processing
word sense disambiguation; semantics; multilinguality
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
With More Contexts Comes Better Performance: Contextualized Sense Embeddings for All-Round Word Sense Disambiguation / Scarlini, Bianca; Pasini, Tommaso; Navigli, Roberto. - (2020), pp. 3528-3539. (Intervento presentato al convegno The 2020 Conference on Empirical Methods in Natural Language Processing tenutosi a Online) [10.18653/v1/2020.emnlp-main.285].
File allegati a questo prodotto
File Dimensione Formato  
Scarlini_With-More-Contexts_2020.pdf

accesso aperto

Note: https://www.aclweb.org/anthology/2020.emnlp-main.285
Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Creative commons
Dimensione 405.72 kB
Formato Adobe PDF
405.72 kB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1500017
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 72
  • ???jsp.display-item.citation.isi??? 35
social impact