Abstract Meaning Representation (AMR) is a Semantic Parsing formalism that aims at providing a semantic graph abstraction representing a given text. Current approaches are based on autoregressive language models such as BART or T5, fine-tuned through Teacher Forcing to obtain a linearized version of the AMR graph from a sentence. In this paper, we present LeakDistill, a model and method that explores a modification to the Transformer architecture, using structural adapters to explicitly incorporate graph information into the learned representations and improve AMR parsing performance. Our experiments show how, by employing word-to-node alignment to embed graph structural information into the encoder at training time, we can obtain state-of-the-art AMR parsing through self-knowledge distillation, even without the use of additional data.

Incorporating Graph Information in Transformer-based AMR Parsing / Vasylenko, Pavlo; Huguet Cabot, Pere Lluís; Martinez Lorenzo, Abelardo Carlos; Navigli, Roberto. - (2023), pp. 1995-2011. (Intervento presentato al convegno Association for Computational Linguistics tenutosi a Toronto, Canada) [10.18653/v1/2023.findings-acl.125].

Incorporating Graph Information in Transformer-based AMR Parsing

Huguet Cabot, Pere Lluís
;
Martinez Lorenzo, Abelardo Carlos
;
Navigli, Roberto
2023

Abstract

Abstract Meaning Representation (AMR) is a Semantic Parsing formalism that aims at providing a semantic graph abstraction representing a given text. Current approaches are based on autoregressive language models such as BART or T5, fine-tuned through Teacher Forcing to obtain a linearized version of the AMR graph from a sentence. In this paper, we present LeakDistill, a model and method that explores a modification to the Transformer architecture, using structural adapters to explicitly incorporate graph information into the learned representations and improve AMR parsing performance. Our experiments show how, by employing word-to-node alignment to embed graph structural information into the encoder at training time, we can obtain state-of-the-art AMR parsing through self-knowledge distillation, even without the use of additional data.
2023
Association for Computational Linguistics
Semantic Parsing; AMR; AMR parsing
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Incorporating Graph Information in Transformer-based AMR Parsing / Vasylenko, Pavlo; Huguet Cabot, Pere Lluís; Martinez Lorenzo, Abelardo Carlos; Navigli, Roberto. - (2023), pp. 1995-2011. (Intervento presentato al convegno Association for Computational Linguistics tenutosi a Toronto, Canada) [10.18653/v1/2023.findings-acl.125].
File allegati a questo prodotto
File Dimensione Formato  
Vasylenko_Incorporating-Graph_2023.pdf

accesso aperto

Note: https://aclanthology.org/2023.findings-acl.125.pdf
Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Creative commons
Dimensione 768.19 kB
Formato Adobe PDF
768.19 kB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1688044
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 8
  • ???jsp.display-item.citation.isi??? ND
social impact