Understanding textual description to generate code seems to be an achieved capability of instruction-following Large Language Models (LLMs) in zero-shot scenario. However, there is a severe possibility that this translation ability may be influenced by having seen target textual descriptions and the related code. This effect is known as Data Contamination.In this study, we investigate the impact of Data Contamination on the performance of GPT-3.5 in the Text-to-SQL code-generating tasks. Hence, we introduce a novel method to detect Data Contamination in GPTs and examine GPT-3.5’s Text-to-SQL performances using the known Spider Dataset and our new unfamiliar dataset Termite. Furthermore, we analyze GPT-3.5’s efficacy on databases with modified information via an adversarial table disconnection (ATD) approach, complicating Text-to-SQL tasks by removing structural pieces of information from the database. Our results indicate a significant performance drop in GPT-3.5 on the unfamiliar Termite dataset, even with ATD modifications, highlighting the effect of Data Contamination on LLMs in Text-to-SQL translation tasks.

Investigating the Impact of Data Contamination of Large Language Models in Text-to-SQL Translation / Ranaldi, F; Ruzzetti, Es; Onorati, D; Ranaldi, L; Giannone, C; Favalli, A; Romagnoli, R; Zanzotto, Fm. - (2024), pp. 13909-13920. (Intervento presentato al convegno 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024) tenutosi a Bangkok; Thailand) [10.18653/v1/2024.findings-acl.827].

Investigating the Impact of Data Contamination of Large Language Models in Text-to-SQL Translation

Onorati, D;
2024

Abstract

Understanding textual description to generate code seems to be an achieved capability of instruction-following Large Language Models (LLMs) in zero-shot scenario. However, there is a severe possibility that this translation ability may be influenced by having seen target textual descriptions and the related code. This effect is known as Data Contamination.In this study, we investigate the impact of Data Contamination on the performance of GPT-3.5 in the Text-to-SQL code-generating tasks. Hence, we introduce a novel method to detect Data Contamination in GPTs and examine GPT-3.5’s Text-to-SQL performances using the known Spider Dataset and our new unfamiliar dataset Termite. Furthermore, we analyze GPT-3.5’s efficacy on databases with modified information via an adversarial table disconnection (ATD) approach, complicating Text-to-SQL tasks by removing structural pieces of information from the database. Our results indicate a significant performance drop in GPT-3.5 on the unfamiliar Termite dataset, even with ATD modifications, highlighting the effect of Data Contamination on LLMs in Text-to-SQL translation tasks.
2024
62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024)
Natural Language Processing; LLMs; Text-to-SQL; Data Contamination
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Investigating the Impact of Data Contamination of Large Language Models in Text-to-SQL Translation / Ranaldi, F; Ruzzetti, Es; Onorati, D; Ranaldi, L; Giannone, C; Favalli, A; Romagnoli, R; Zanzotto, Fm. - (2024), pp. 13909-13920. (Intervento presentato al convegno 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024) tenutosi a Bangkok; Thailand) [10.18653/v1/2024.findings-acl.827].
File allegati a questo prodotto
File Dimensione Formato  
Ranaldi_Investigating_2024.pdf

accesso aperto

Note: DOI: 10.18653/v1/2024.findings-acl.827 - PDF: https://aclanthology.org/2024.findings-acl.827.pdf
Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 189.25 kB
Formato Adobe PDF
189.25 kB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1727083
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact