Existing methods for interpreting predictions from Graph Neural Networks (GNNs) have primarily focused on generating subgraphs that are especially relevant for a particular prediction. However, such methods do not provide a clear opportunity for recourse: given a prediction, we want to understand how the prediction can be changed in order to achieve a more desirable outcome. In this work, we propose a method for generating counterfac- tual (CF) explanations for GNNs: the mini- mal perturbation to the input (graph) data such that the prediction changes. Using only edge deletions, we find that our method, CF- GNNExplainer, can generate CF explana- tions for the majority of instances across three widely used datasets for GNN explanations, while removing less than 3 edges on average, with at least 94% accuracy. This indicates that CF-GNNExplainer primarily removes edges that are crucial for the original predic- tions, resulting in minimal CF explanations.

CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks / Lucic, Ana; ter Hoeve, Maartje; Tolomei, Gabriele; de Rijke, Maarten; Silvestri, Fabrizio. - (2022). (Intervento presentato al convegno International Conference on Artificial Intelligence and Statistics tenutosi a Virtual Conference).

CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks

Gabriele Tolomei
Writing – Original Draft Preparation
;
Fabrizio Silvestri
Writing – Original Draft Preparation
2022

Abstract

Existing methods for interpreting predictions from Graph Neural Networks (GNNs) have primarily focused on generating subgraphs that are especially relevant for a particular prediction. However, such methods do not provide a clear opportunity for recourse: given a prediction, we want to understand how the prediction can be changed in order to achieve a more desirable outcome. In this work, we propose a method for generating counterfac- tual (CF) explanations for GNNs: the mini- mal perturbation to the input (graph) data such that the prediction changes. Using only edge deletions, we find that our method, CF- GNNExplainer, can generate CF explana- tions for the majority of instances across three widely used datasets for GNN explanations, while removing less than 3 edges on average, with at least 94% accuracy. This indicates that CF-GNNExplainer primarily removes edges that are crucial for the original predic- tions, resulting in minimal CF explanations.
2022
International Conference on Artificial Intelligence and Statistics
graph neural networks, explanation, counterfactual explanation, XAI
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks / Lucic, Ana; ter Hoeve, Maartje; Tolomei, Gabriele; de Rijke, Maarten; Silvestri, Fabrizio. - (2022). (Intervento presentato al convegno International Conference on Artificial Intelligence and Statistics tenutosi a Virtual Conference).
File allegati a questo prodotto
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1627349
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 29
  • ???jsp.display-item.citation.isi??? 2
social impact