We consider dense, associative neural-networks trained by a teacher (i.e., with supervision) and we investigate their computational capabilities analytically, via statistical-mechanics tools, and numerically, via Monte Carlo simulations. In particular, we obtain a phase diagram which summarizes their performance as a function of the control parameters (e.g., quality and quantity of the training dataset, network storage, noise), that is valid in the limit of large network-size and structureless datasets. We also numerically test the learning, storing and retrieval capabilities of these networks on structured datasets such as MNist and Fashion MNist. As technical remarks, on the analytic side, we extend Guerra’s interpolation to tackle the non-Gaussian distributions involved in the post-synaptic potentials while, on the computational side, we insert Plefka’s approximation in the Monte Carlo scheme, to speed up the evaluation of the synaptic tensors, overall obtaining a novel and broad approach to investigate supervised learning in neural networks, beyond the shallow limit.

Dense Hebbian neural networks: A replica symmetric picture of supervised learning / Agliari, Elena; Albanese, Linda; Alemanno, Francesco; Alessandrelli, Andrea; Barra, Adriano; Giannotti, Fosca; Lotito, Daniele; Pedreschi, Dino. - In: PHYSICA. A. - ISSN 0378-4371. - 626:(2023), p. 129076. [10.1016/j.physa.2023.129076]

Dense Hebbian neural networks: A replica symmetric picture of supervised learning

Elena Agliari;Adriano Barra;
2023

Abstract

We consider dense, associative neural-networks trained by a teacher (i.e., with supervision) and we investigate their computational capabilities analytically, via statistical-mechanics tools, and numerically, via Monte Carlo simulations. In particular, we obtain a phase diagram which summarizes their performance as a function of the control parameters (e.g., quality and quantity of the training dataset, network storage, noise), that is valid in the limit of large network-size and structureless datasets. We also numerically test the learning, storing and retrieval capabilities of these networks on structured datasets such as MNist and Fashion MNist. As technical remarks, on the analytic side, we extend Guerra’s interpolation to tackle the non-Gaussian distributions involved in the post-synaptic potentials while, on the computational side, we insert Plefka’s approximation in the Monte Carlo scheme, to speed up the evaluation of the synaptic tensors, overall obtaining a novel and broad approach to investigate supervised learning in neural networks, beyond the shallow limit.
2023
Spin glasses; Hebbian learning; Dense networks
01 Pubblicazione su rivista::01a Articolo in rivista
Dense Hebbian neural networks: A replica symmetric picture of supervised learning / Agliari, Elena; Albanese, Linda; Alemanno, Francesco; Alessandrelli, Andrea; Barra, Adriano; Giannotti, Fosca; Lotito, Daniele; Pedreschi, Dino. - In: PHYSICA. A. - ISSN 0378-4371. - 626:(2023), p. 129076. [10.1016/j.physa.2023.129076]
File allegati a questo prodotto
File Dimensione Formato  
Agliari_Dense-Hebbian_2023.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 2.15 MB
Formato Adobe PDF
2.15 MB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1689059
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? ND
social impact