Deep neural networks (DNNs) are widely used in critical applications such as autonomous systems, healthcare, and finance. However, their increasing deployment raises security concerns, particularly against adversarial threats like Trojan attacks. These attacks embed hidden triggers during training, causing misclassifications when activated while maintaining normal performance otherwise, making them difficult to detect. We propose a novel graph-based detection framework that models neuron activations as graph structures, capturing topological anomalies indicative of Trojan behavior. We leverage Graph Neural Networks (GNNs) to identify these structural deviations effectively. Additionally, we integrate Explainable AI (XAI) techniques to enhance interpretability. This not only improves trust and transparency in our framework but also aids in understanding the fundamental characteristics of backdoored networks, potentially guiding the development of more robust defense mechanisms.
Topology-Driven Explainable GNNs for Trojan Detection in Deep Learning / Ceccaroni, Riccardo; Brutti, Pierpaolo. - (2025). ( International Joint Conference on Neural Networks Rome; Italy ) [10.1109/IJCNN64981.2025.11227820].
Topology-Driven Explainable GNNs for Trojan Detection in Deep Learning
Riccardo Ceccaroni
;Pierpaolo Brutti
2025
Abstract
Deep neural networks (DNNs) are widely used in critical applications such as autonomous systems, healthcare, and finance. However, their increasing deployment raises security concerns, particularly against adversarial threats like Trojan attacks. These attacks embed hidden triggers during training, causing misclassifications when activated while maintaining normal performance otherwise, making them difficult to detect. We propose a novel graph-based detection framework that models neuron activations as graph structures, capturing topological anomalies indicative of Trojan behavior. We leverage Graph Neural Networks (GNNs) to identify these structural deviations effectively. Additionally, we integrate Explainable AI (XAI) techniques to enhance interpretability. This not only improves trust and transparency in our framework but also aids in understanding the fundamental characteristics of backdoored networks, potentially guiding the development of more robust defense mechanisms.| File | Dimensione | Formato | |
|---|---|---|---|
|
Ceccaroni_Topology-driven_2025.pdf
solo gestori archivio
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
545.37 kB
Formato
Adobe PDF
|
545.37 kB | Adobe PDF | Contatta l'autore |
|
Ceccaroni_Topology-drive_copertina_2025.pdf
solo gestori archivio
Note: quarta di copertina
Tipologia:
Altro materiale allegato
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
611.9 kB
Formato
Adobe PDF
|
611.9 kB | Adobe PDF | Contatta l'autore |
|
Ceccaroni_Topology-driven_frontespizio_2025.pdf
solo gestori archivio
Note: frontespizio
Tipologia:
Altro materiale allegato
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
862.96 kB
Formato
Adobe PDF
|
862.96 kB | Adobe PDF | Contatta l'autore |
|
Ceccaroni_Topology-driven_indice_2025.pdf
solo gestori archivio
Note: indice
Tipologia:
Altro materiale allegato
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
870.94 kB
Formato
Adobe PDF
|
870.94 kB | Adobe PDF | Contatta l'autore |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


