Deep learning has revolutionized performance across diverse applications but is often criticized for its opaque nature and computational complexity, complicating efforts to understand and trust model predictions. Explainability techniques have emerged to address these issues, enhancing model transparency and user trust. Simultaneously, efficiency techniques such as Knowledge Distillation offer solutions to streamline complex (teacher) models into more efficient (student) ones while preserving their predictive capabilities. Building on those research efforts, this paper presents a general framework that is able to improve well-known distillation techniques through explainability methods boosting both model interpretability and accuracy of shallow student architectures. More in detail, by incorporating a novel loss function that aligns the explainability maps of teacher and student models, our method refines the distillation process, leading to more accurate and interpretable predictions. Our approach not only improves the performance of distilled models but also demonstrates the effective integration of explainable insights into distillation frameworks, as validated by multiple experiments on a publicly available benchmark dataset.
Explaining the Inexplicable: An Explainable AI Approach to Boost Knowledge Transfer / Conforti, Pietro; Papa, Lorenzo; Amerini, Irene; Russo, Paolo. - (2026), pp. 660-670. ( International Conference on Computer Vision Theory and Applications (VISAPP 2026) Marbella, Spain ) [10.5220/0014588800004084].
Explaining the Inexplicable: An Explainable AI Approach to Boost Knowledge Transfer
Conforti, Pietro;Amerini, Irene;
2026
Abstract
Deep learning has revolutionized performance across diverse applications but is often criticized for its opaque nature and computational complexity, complicating efforts to understand and trust model predictions. Explainability techniques have emerged to address these issues, enhancing model transparency and user trust. Simultaneously, efficiency techniques such as Knowledge Distillation offer solutions to streamline complex (teacher) models into more efficient (student) ones while preserving their predictive capabilities. Building on those research efforts, this paper presents a general framework that is able to improve well-known distillation techniques through explainability methods boosting both model interpretability and accuracy of shallow student architectures. More in detail, by incorporating a novel loss function that aligns the explainability maps of teacher and student models, our method refines the distillation process, leading to more accurate and interpretable predictions. Our approach not only improves the performance of distilled models but also demonstrates the effective integration of explainable insights into distillation frameworks, as validated by multiple experiments on a publicly available benchmark dataset.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


