Recently, a daily routine for associative neural networks has been proposed: the network Hebbian-learns during the awake state (thus behaving as a standard Hopfield model), then, during its sleep state, it consolidates pure patterns and removes spurious ones, optimizing information storage: this forces the synaptic matrix to collapse to the projector one (ultimately approaching the Kanter-Sompolinsky model), allowing for the maximal critical capacity (for symmetric interactions). So far, this emerging picture (as well as the bulk of papers on unlearning techniques) was supported solely by non-rigorous routes, e.g. replica-trick analysis and numerical simulations, while here we rely on Guerra's interpolation techniques and we extend the generalized stochastic stability approach to the case. Focusing on the replica-symmetric scenario (where the previous investigations lie), the former picture is entirely confirmed. Further, we develop a fluctuation analysis to check where ergodicity is broken (an analysis absent in previous investigations). Remarkably, we find that, as long as the network is awake, ergodicity is bounded by the Amit-Gutfreund-Sompolinsky critical line (as it should), but, as the network sleeps, spin-glass states are destroyed and both the retrieval and the ergodic region get wider. Thus, after a whole sleeping session, the solely surviving regions are the retrieval and ergodic ones, in such a way that the network achieves a perfect retrieval regime.
Dreaming neural networks: Rigorous results / Agliari, E.; Alemanno, F.; Barra, A.; Fachechi, A.. - In: JOURNAL OF STATISTICAL MECHANICS: THEORY AND EXPERIMENT. - ISSN 1742-5468. - 2019:8(2019), p. 083503. [10.1088/1742-5468/ab371d]
Dreaming neural networks: Rigorous results
Agliari E.;Barra A.;Fachechi A.
2019
Abstract
Recently, a daily routine for associative neural networks has been proposed: the network Hebbian-learns during the awake state (thus behaving as a standard Hopfield model), then, during its sleep state, it consolidates pure patterns and removes spurious ones, optimizing information storage: this forces the synaptic matrix to collapse to the projector one (ultimately approaching the Kanter-Sompolinsky model), allowing for the maximal critical capacity (for symmetric interactions). So far, this emerging picture (as well as the bulk of papers on unlearning techniques) was supported solely by non-rigorous routes, e.g. replica-trick analysis and numerical simulations, while here we rely on Guerra's interpolation techniques and we extend the generalized stochastic stability approach to the case. Focusing on the replica-symmetric scenario (where the previous investigations lie), the former picture is entirely confirmed. Further, we develop a fluctuation analysis to check where ergodicity is broken (an analysis absent in previous investigations). Remarkably, we find that, as long as the network is awake, ergodicity is bounded by the Amit-Gutfreund-Sompolinsky critical line (as it should), but, as the network sleeps, spin-glass states are destroyed and both the retrieval and the ergodic region get wider. Thus, after a whole sleeping session, the solely surviving regions are the retrieval and ergodic ones, in such a way that the network achieves a perfect retrieval regime.File | Dimensione | Formato | |
---|---|---|---|
Agliari_Dreaming-neural_2019.pdf
solo gestori archivio
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
1.37 MB
Formato
Adobe PDF
|
1.37 MB | Adobe PDF | Contatta l'autore |
Agliari_preprint_Dreaming-neural_2019.pdf
accesso aperto
Tipologia:
Documento in Pre-print (manoscritto inviato all'editore, precedente alla peer review)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
1.06 MB
Formato
Adobe PDF
|
1.06 MB | Adobe PDF |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.