Lexical ambiguity poses one of the greatest challenges in the field of Machine Translation. Over the last few decades, multiple efforts have been undertaken to investigate incorrect translations caused by the polysemous nature of words. Within this body of research, some studies have posited that models pick up semantic biases existing in the training data, thus producing translation errors. In this paper, we present DiBiMT, the first entirely manually-curated evaluation benchmark which enables an extensive study of semantic biases in Machine Translation of nominal and verbal words in five different language combinations, namely, English and one or other of the following languages: Chinese, German, Italian, Russian and Spanish. Furthermore, we test state-of-the-art Machine Translation systems, both commercial and non-commercial ones, against our new test bed and provide a thorough statistical and linguistic analysis of the results. We release DiBiMT at https://nlp.uniroma1.it/dibimt as a closed benchmark with a public leaderboard.

DiBiMT: A Novel Benchmark for Measuring Word Sense Disambiguation Biases in Machine Translation / Campolungo, Niccolò; Martelli, Federico; Saina, Francesco; Navigli, Roberto. - (2022), pp. 4331-4352. (Intervento presentato al convegno Association for Computational Linguistics tenutosi a Dublin; Ireland) [10.18653/v1/2022.acl-long.298].

DiBiMT: A Novel Benchmark for Measuring Word Sense Disambiguation Biases in Machine Translation

Campolungo, Niccolò
Primo
;
Martelli, Federico
;
Navigli, Roberto
2022

Abstract

Lexical ambiguity poses one of the greatest challenges in the field of Machine Translation. Over the last few decades, multiple efforts have been undertaken to investigate incorrect translations caused by the polysemous nature of words. Within this body of research, some studies have posited that models pick up semantic biases existing in the training data, thus producing translation errors. In this paper, we present DiBiMT, the first entirely manually-curated evaluation benchmark which enables an extensive study of semantic biases in Machine Translation of nominal and verbal words in five different language combinations, namely, English and one or other of the following languages: Chinese, German, Italian, Russian and Spanish. Furthermore, we test state-of-the-art Machine Translation systems, both commercial and non-commercial ones, against our new test bed and provide a thorough statistical and linguistic analysis of the results. We release DiBiMT at https://nlp.uniroma1.it/dibimt as a closed benchmark with a public leaderboard.
2022
Association for Computational Linguistics
DiBiMT; machine translation; disambiguation bias; semantic bias; bias
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
DiBiMT: A Novel Benchmark for Measuring Word Sense Disambiguation Biases in Machine Translation / Campolungo, Niccolò; Martelli, Federico; Saina, Francesco; Navigli, Roberto. - (2022), pp. 4331-4352. (Intervento presentato al convegno Association for Computational Linguistics tenutosi a Dublin; Ireland) [10.18653/v1/2022.acl-long.298].
File allegati a questo prodotto
File Dimensione Formato  
Campolungo_DIBIMT_2022.pdf

accesso aperto

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Creative commons
Dimensione 660.22 kB
Formato Adobe PDF
660.22 kB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1640533
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 19
  • ???jsp.display-item.citation.isi??? 5
social impact