Deep neural networks have been pivotal in driving AI advancements over the past decade, revolutionizing domains such as gaming, biology, autonomous systems, and voice and text assistants. Despite their impact, these networks are often regarded as black-box models due to their intricate structures and the absence of explanations for their decisions. This opacity poses a significant challenge to AI systems' wider adoption and trustworthiness. This thesis addresses this issue by contributing to the field of eXplainable AI, focusing on enhancing the interpretability of deep neural networks. The core contributions lie in introducing novel techniques aimed at making these networks more interpretable by leveraging an analysis of their inner workings. Specifically, the contributions are threefold. Firstly, the thesis introduces designs for self-explanatory deep neural networks, such as the integration of external memory for interpretability purposes and the usage of prototype and constraint-based layers across several domains. These proposed architectures are specifically designed to preserve most of the black-box networks, thereby maintaining or improving their performance. Secondly, this research delves into novel investigations on neurons within trained deep neural networks, shedding light on overlooked phenomena related to their activation values. Lastly, the thesis conducts an analysis of the application of explanatory techniques in the field of visual analytics, exploring the maturity of their adoption and the potential of these systems to convey explanations to users effectively. In summary, this thesis contributes to the growing field of Explainable AI by proposing intrinsic techniques to enhance the interpretability of deep neural networks. By mitigating the opacity issue of deep neural networks and applying them to several different applications, the research aims to foster trust in AI systems and facilitate their wider adoption across several applications.

Explaining deep neural networks by leveraging intrinsic methods / LA ROSA, Biagio. - (2024 May 07).

Explaining deep neural networks by leveraging intrinsic methods

LA ROSA, BIAGIO
07/05/2024

Abstract

Deep neural networks have been pivotal in driving AI advancements over the past decade, revolutionizing domains such as gaming, biology, autonomous systems, and voice and text assistants. Despite their impact, these networks are often regarded as black-box models due to their intricate structures and the absence of explanations for their decisions. This opacity poses a significant challenge to AI systems' wider adoption and trustworthiness. This thesis addresses this issue by contributing to the field of eXplainable AI, focusing on enhancing the interpretability of deep neural networks. The core contributions lie in introducing novel techniques aimed at making these networks more interpretable by leveraging an analysis of their inner workings. Specifically, the contributions are threefold. Firstly, the thesis introduces designs for self-explanatory deep neural networks, such as the integration of external memory for interpretability purposes and the usage of prototype and constraint-based layers across several domains. These proposed architectures are specifically designed to preserve most of the black-box networks, thereby maintaining or improving their performance. Secondly, this research delves into novel investigations on neurons within trained deep neural networks, shedding light on overlooked phenomena related to their activation values. Lastly, the thesis conducts an analysis of the application of explanatory techniques in the field of visual analytics, exploring the maturity of their adoption and the potential of these systems to convey explanations to users effectively. In summary, this thesis contributes to the growing field of Explainable AI by proposing intrinsic techniques to enhance the interpretability of deep neural networks. By mitigating the opacity issue of deep neural networks and applying them to several different applications, the research aims to foster trust in AI systems and facilitate their wider adoption across several applications.
7-mag-2024
File allegati a questo prodotto
File Dimensione Formato  
Tasi_dottorato_LaRosa.pdf

accesso aperto

Note: Tesi completa
Tipologia: Tesi di dottorato
Licenza: Creative commons
Dimensione 18.57 MB
Formato Adobe PDF
18.57 MB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1715783
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact