Lexical ambiguity is one of the many challenging linguistic phenomena involved in translation, i.e., translating an ambiguous word with its correct sense. In this respect, previous work has shown that the translation quality of neural machine translation systems can be improved by explicitly modeling the senses of ambiguous words. Recently, several evaluation test sets have been proposed to measure the word sense disambiguation (WSD) capability of machine translation systems. However, to date, these evaluation test sets do not include any training data that would provide a fair setup measuring the sense distributions present within the training data itself. In this paper, we present an evaluation benchmark on WSD for machine translation for 10 language pairs, comprising training data with known sense distributions. Our approach for the construction of the benchmark builds upon the wide-coverage multilingual sense inventory of BabelNet, the multilingual neural parsing pipeline TurkuNLP, and the OPUS collection of translated texts from the web. The test suite is available at http://github.com/Helsinki-NLP/MuCoW.

An Evaluation Benchmark for Testing the Word Sense Disambiguation Capabilities of Machine Translation Systems / Raganato, Alessandro; Scherrer, Yves; Tiedemann, Jörg. - (2020), pp. 3668-3675. ( 12th International Conference on Language Resources and Evaluation, LREC 2020 Marseille; France ).

An Evaluation Benchmark for Testing the Word Sense Disambiguation Capabilities of Machine Translation Systems

Raganato Alessandro
;
2020

Abstract

Lexical ambiguity is one of the many challenging linguistic phenomena involved in translation, i.e., translating an ambiguous word with its correct sense. In this respect, previous work has shown that the translation quality of neural machine translation systems can be improved by explicitly modeling the senses of ambiguous words. Recently, several evaluation test sets have been proposed to measure the word sense disambiguation (WSD) capability of machine translation systems. However, to date, these evaluation test sets do not include any training data that would provide a fair setup measuring the sense distributions present within the training data itself. In this paper, we present an evaluation benchmark on WSD for machine translation for 10 language pairs, comprising training data with known sense distributions. Our approach for the construction of the benchmark builds upon the wide-coverage multilingual sense inventory of BabelNet, the multilingual neural parsing pipeline TurkuNLP, and the OPUS collection of translated texts from the web. The test suite is available at http://github.com/Helsinki-NLP/MuCoW.
2020
12th International Conference on Language Resources and Evaluation, LREC 2020
lexical ambiguity; machine translation; word sense disambiguation
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
An Evaluation Benchmark for Testing the Word Sense Disambiguation Capabilities of Machine Translation Systems / Raganato, Alessandro; Scherrer, Yves; Tiedemann, Jörg. - (2020), pp. 3668-3675. ( 12th International Conference on Language Resources and Evaluation, LREC 2020 Marseille; France ).
File allegati a questo prodotto
File Dimensione Formato  
Raganato_An-evaluation-benchmark_2020.pdf

accesso aperto

Note: https://aclanthology.org/2020.lrec-1.452/
Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 340.67 kB
Formato Adobe PDF
340.67 kB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1553735
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 13
  • ???jsp.display-item.citation.isi??? 6
social impact