Entity resolution (ER) aims at identifying record pairs that refer to the same real-world entity. Recent works have focused on deep learning (DL) techniques, to solve this problem. While such works have brought tremendous enhancements in terms of effectiveness in solving the ER problem, understanding their matching predictions is still a challenge, because of the intrinsic opaqueness of DL based solutions. Interpreting and trusting the predictions made by ER systems is crucial for humans in order to employ such methods in decision making pipelines. We demonstrate certem an explanation system for ER based on certa, a recently introduced explainability framework for ER, that is able to provide both saliency explana- tions, which associate each attribute with a saliency score, and counterfactual explanations, which provide examples of values that can flip a prediction. In this demonstration we will showcase how certem can be effectively employed to better understand and debug the behavior of state-of-the-art DL based ER systems on data from publicly available ER benchmarks.

CERTEM: Explaining and Debugging Black-box Entity Resolution Systems with CERTA / Teofili, Tommaso; Firmani, Donatella; Koudas, Nick; Merialdo, Paolo; Srivastava, Divesh. - In: PROCEEDINGS OF THE VLDB ENDOWMENT. - ISSN 2150-8097. - (2022), pp. 1-4.

CERTEM: Explaining and Debugging Black-box Entity Resolution Systems with CERTA

Donatella Firmani;
2022

Abstract

Entity resolution (ER) aims at identifying record pairs that refer to the same real-world entity. Recent works have focused on deep learning (DL) techniques, to solve this problem. While such works have brought tremendous enhancements in terms of effectiveness in solving the ER problem, understanding their matching predictions is still a challenge, because of the intrinsic opaqueness of DL based solutions. Interpreting and trusting the predictions made by ER systems is crucial for humans in order to employ such methods in decision making pipelines. We demonstrate certem an explanation system for ER based on certa, a recently introduced explainability framework for ER, that is able to provide both saliency explana- tions, which associate each attribute with a saliency score, and counterfactual explanations, which provide examples of values that can flip a prediction. In this demonstration we will showcase how certem can be effectively employed to better understand and debug the behavior of state-of-the-art DL based ER systems on data from publicly available ER benchmarks.
2022
explainable AI; data cleaning; entity resolution
01 Pubblicazione su rivista::01a Articolo in rivista
CERTEM: Explaining and Debugging Black-box Entity Resolution Systems with CERTA / Teofili, Tommaso; Firmani, Donatella; Koudas, Nick; Merialdo, Paolo; Srivastava, Divesh. - In: PROCEEDINGS OF THE VLDB ENDOWMENT. - ISSN 2150-8097. - (2022), pp. 1-4.
File allegati a questo prodotto
File Dimensione Formato  
Teofili_CERTEM_2022.pdf

accesso aperto

Tipologia: Documento in Post-print (versione successiva alla peer review e accettata per la pubblicazione)
Licenza: Creative commons
Dimensione 1.27 MB
Formato Adobe PDF
1.27 MB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1642697
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? 0
social impact