Deep neural networks (DNNs) are widely used in critical applications such as autonomous systems, healthcare, and finance. However, their increasing deployment raises security concerns, particularly against adversarial threats like Trojan attacks. These attacks embed hidden triggers during training, causing misclassifications when activated while maintaining normal performance otherwise, making them difficult to detect. We propose a novel graph-based detection framework that models neuron activations as graph structures, capturing topological anomalies indicative of Trojan behavior. We leverage Graph Neural Networks (GNNs) to identify these structural deviations effectively. Additionally, we integrate Explainable AI (XAI) techniques to enhance interpretability. This not only improves trust and transparency in our framework but also aids in understanding the fundamental characteristics of backdoored networks, potentially guiding the development of more robust defense mechanisms.

Topology-Driven Explainable GNNs for Trojan Detection in Deep Learning / Ceccaroni, Riccardo; Brutti, Pierpaolo. - (2025). ( International Joint Conference on Neural Networks Rome; Italy ) [10.1109/IJCNN64981.2025.11227820].

Topology-Driven Explainable GNNs for Trojan Detection in Deep Learning

Riccardo Ceccaroni
;
Pierpaolo Brutti
2025

Abstract

Deep neural networks (DNNs) are widely used in critical applications such as autonomous systems, healthcare, and finance. However, their increasing deployment raises security concerns, particularly against adversarial threats like Trojan attacks. These attacks embed hidden triggers during training, causing misclassifications when activated while maintaining normal performance otherwise, making them difficult to detect. We propose a novel graph-based detection framework that models neuron activations as graph structures, capturing topological anomalies indicative of Trojan behavior. We leverage Graph Neural Networks (GNNs) to identify these structural deviations effectively. Additionally, we integrate Explainable AI (XAI) techniques to enhance interpretability. This not only improves trust and transparency in our framework but also aids in understanding the fundamental characteristics of backdoored networks, potentially guiding the development of more robust defense mechanisms.
2025
International Joint Conference on Neural Networks
Trojan attack detection; Explainable AI; Topological data analysis; Machine learning security
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Topology-Driven Explainable GNNs for Trojan Detection in Deep Learning / Ceccaroni, Riccardo; Brutti, Pierpaolo. - (2025). ( International Joint Conference on Neural Networks Rome; Italy ) [10.1109/IJCNN64981.2025.11227820].
File allegati a questo prodotto
File Dimensione Formato  
Ceccaroni_Topology-driven_2025.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 545.37 kB
Formato Adobe PDF
545.37 kB Adobe PDF   Contatta l'autore
Ceccaroni_Topology-drive_copertina_2025.pdf

solo gestori archivio

Note: quarta di copertina
Tipologia: Altro materiale allegato
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 611.9 kB
Formato Adobe PDF
611.9 kB Adobe PDF   Contatta l'autore
Ceccaroni_Topology-driven_frontespizio_2025.pdf

solo gestori archivio

Note: frontespizio
Tipologia: Altro materiale allegato
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 862.96 kB
Formato Adobe PDF
862.96 kB Adobe PDF   Contatta l'autore
Ceccaroni_Topology-driven_indice_2025.pdf

solo gestori archivio

Note: indice
Tipologia: Altro materiale allegato
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 870.94 kB
Formato Adobe PDF
870.94 kB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1748965
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact