Two of the most impressive features of biological neural networks are their high energy efficiency and their ability to continuously adapt to varying inputs. On the contrary, the amount of power required to train top-performing deep learning models rises as they become more complex. This is the main reason for the increasing research interest in spiking neural networks, which mimic the functioning of the human brain achieving similar performances to artificial neural networks, but with much lower energy costs. However, even this type of network is not provided with the ability to incrementally learn new tasks, with the main obstacle being catastrophic forgetting. This paper investigates memory replay as a strategy to mitigate catastrophic forgetting in spiking neural networks. Experiments are conducted on the MNIST-split dataset in both class-incremental learning and task-free continual learning scenarios.

Memory Replay For Continual Learning With Spiking Neural Networks / Proietti, Michela; Ragno, Alessio; Capobianco, Roberto. - (2023), pp. 1-6. (Intervento presentato al convegno 2023 IEEE 33rd International Workshop on Machine Learning for Signal Processing (MLSP) tenutosi a Rome; Italy) [10.1109/MLSP55844.2023.10285911].

Memory Replay For Continual Learning With Spiking Neural Networks

Proietti, Michela
;
Ragno, Alessio;Capobianco, Roberto
2023

Abstract

Two of the most impressive features of biological neural networks are their high energy efficiency and their ability to continuously adapt to varying inputs. On the contrary, the amount of power required to train top-performing deep learning models rises as they become more complex. This is the main reason for the increasing research interest in spiking neural networks, which mimic the functioning of the human brain achieving similar performances to artificial neural networks, but with much lower energy costs. However, even this type of network is not provided with the ability to incrementally learn new tasks, with the main obstacle being catastrophic forgetting. This paper investigates memory replay as a strategy to mitigate catastrophic forgetting in spiking neural networks. Experiments are conducted on the MNIST-split dataset in both class-incremental learning and task-free continual learning scenarios.
2023
2023 IEEE 33rd International Workshop on Machine Learning for Signal Processing (MLSP)
spiking neural networks; continual learning; memory replay
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Memory Replay For Continual Learning With Spiking Neural Networks / Proietti, Michela; Ragno, Alessio; Capobianco, Roberto. - (2023), pp. 1-6. (Intervento presentato al convegno 2023 IEEE 33rd International Workshop on Machine Learning for Signal Processing (MLSP) tenutosi a Rome; Italy) [10.1109/MLSP55844.2023.10285911].
File allegati a questo prodotto
File Dimensione Formato  
Proietti_Memory_2023.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 697.59 kB
Formato Adobe PDF
697.59 kB Adobe PDF   Contatta l'autore
Proietti_postprint_Memory_2023.pdf

accesso aperto

Note: DOI:10.1109/MLSP55844.2023.10285911
Tipologia: Documento in Post-print (versione successiva alla peer review e accettata per la pubblicazione)
Licenza: Creative commons
Dimensione 629.63 kB
Formato Adobe PDF
629.63 kB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1690926
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact