: Hebb's learning traces its origin in Pavlov's classical conditioning; however, while the former has been extensively modeled in the past decades (e.g., by the Hopfield model and countless variations on theme), as for the latter, modeling has remained largely unaddressed so far. Furthermore, a mathematical bridge connecting these two pillars is totally lacking. The main difficulty toward this goal lies in the intrinsically different scales of the information involved: Pavlov's theory is about correlations between concepts that are (dynamically) stored in the synaptic matrix as exemplified by the celebrated experiment starring a dog and a ringing bell; conversely, Hebb's theory is about correlations between pairs of neurons as summarized by the famous statement that neurons that fire together wire together. In this letter, we rely on stochastic process theory to prove that as long as we keep neurons' and synapses' timescales largely split, Pavlov's mechanism spontaneously takes place and ultimately gives rise to synaptic weights that recover the Hebbian kernel.

From Pavlov Conditioning to Hebb Learning / Agliari, Elena; Aquaro, Miriam; Barra, Adriano; Fachechi, Alberto; Marullo, Chiara. - In: NEURAL COMPUTATION. - ISSN 0899-7667. - 35:5(2023), pp. 930-957. [10.1162/neco_a_01578]

From Pavlov Conditioning to Hebb Learning

Agliari, Elena
Primo
;
Aquaro, Miriam;Barra, Adriano
;
Fachechi, Alberto;Marullo, Chiara
2023

Abstract

: Hebb's learning traces its origin in Pavlov's classical conditioning; however, while the former has been extensively modeled in the past decades (e.g., by the Hopfield model and countless variations on theme), as for the latter, modeling has remained largely unaddressed so far. Furthermore, a mathematical bridge connecting these two pillars is totally lacking. The main difficulty toward this goal lies in the intrinsically different scales of the information involved: Pavlov's theory is about correlations between concepts that are (dynamically) stored in the synaptic matrix as exemplified by the celebrated experiment starring a dog and a ringing bell; conversely, Hebb's theory is about correlations between pairs of neurons as summarized by the famous statement that neurons that fire together wire together. In this letter, we rely on stochastic process theory to prove that as long as we keep neurons' and synapses' timescales largely split, Pavlov's mechanism spontaneously takes place and ultimately gives rise to synaptic weights that recover the Hebbian kernel.
2023
Neural network; Hebb rule; Pavlov conditioning
01 Pubblicazione su rivista::01f Lettera, Nota
From Pavlov Conditioning to Hebb Learning / Agliari, Elena; Aquaro, Miriam; Barra, Adriano; Fachechi, Alberto; Marullo, Chiara. - In: NEURAL COMPUTATION. - ISSN 0899-7667. - 35:5(2023), pp. 930-957. [10.1162/neco_a_01578]
File allegati a questo prodotto
File Dimensione Formato  
Agliari_From-Pavlov_2023.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 2.01 MB
Formato Adobe PDF
2.01 MB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1678504
Citazioni
  • ???jsp.display-item.citation.pmc??? 1
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 2
social impact