Incorporating novelties into deep learning systems remains a challenging problem. Introducing new information to a machine learning system can interfere with previously stored data and potentially alter the global model paradigm, especially when dealing with non-stationary sources. In such cases, traditional approaches based on validation error minimization offer limited advantages. To address this, we propose a training algorithm inspired by Stuart Kauffman's notion of the Adjacent Possible. This novel training methodology explores new data spaces during the learning phase. It predisposes the neural network to smoothly accept and integrate data sequences with different statistical characteristics than expected. The maximum distance compatible with such inclusion depends on a specific parameter: the sampling temperature used in the explorative phase of the present method. This algorithm, called Dreaming Learning, anticipates potential regime shifts over time, enhancing the neural network's responsiveness to non-stationary events that alter statistical properties. To assess the advantages of this approach, we apply this methodology to unexpected statistical changes in Markov chains and non-stationary dynamics in textual sequences. We demonstrated its ability to improve the auto-correlation of generated textual sequences by ∼29% and enhance the velocity of loss convergence by ∼100% in the case of a paradigm shift in Markov chains.

Dreaming Learning / Londei, Alessandro; Benati, Matteo; Lanzieri, Denise; Loreto, Vittorio. - (2024). (Intervento presentato al convegno 38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024) tenutosi a Vancouver, Canada).

Dreaming Learning

Alessandro Londei
;
Matteo Benati
;
Vittorio Loreto
2024

Abstract

Incorporating novelties into deep learning systems remains a challenging problem. Introducing new information to a machine learning system can interfere with previously stored data and potentially alter the global model paradigm, especially when dealing with non-stationary sources. In such cases, traditional approaches based on validation error minimization offer limited advantages. To address this, we propose a training algorithm inspired by Stuart Kauffman's notion of the Adjacent Possible. This novel training methodology explores new data spaces during the learning phase. It predisposes the neural network to smoothly accept and integrate data sequences with different statistical characteristics than expected. The maximum distance compatible with such inclusion depends on a specific parameter: the sampling temperature used in the explorative phase of the present method. This algorithm, called Dreaming Learning, anticipates potential regime shifts over time, enhancing the neural network's responsiveness to non-stationary events that alter statistical properties. To assess the advantages of this approach, we apply this methodology to unexpected statistical changes in Markov chains and non-stationary dynamics in textual sequences. We demonstrated its ability to improve the auto-correlation of generated textual sequences by ∼29% and enhance the velocity of loss convergence by ∼100% in the case of a paradigm shift in Markov chains.
2024
38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024)
machine learning; adjacent possible; novelty dynamics
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Dreaming Learning / Londei, Alessandro; Benati, Matteo; Lanzieri, Denise; Loreto, Vittorio. - (2024). (Intervento presentato al convegno 38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024) tenutosi a Vancouver, Canada).
File allegati a questo prodotto
File Dimensione Formato  
Londei_Dreaming-Learning_2024.pdf

accesso aperto

Tipologia: Documento in Post-print (versione successiva alla peer review e accettata per la pubblicazione)
Licenza: Creative commons
Dimensione 1.56 MB
Formato Adobe PDF
1.56 MB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1747365
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact