Non-Markovian Reinforcement Learning (RL) tasks present significant challenges, as agents must reason over entire trajectories of state-action pairs to make optimal decisions. A common strategy to address this is through symbolic formalisms, such as Linear Temporal Logic (LTL) or automata, which provide a structured way to express temporally extended objectives. However, these approaches often rely on restrictive assumptions -- such as the availability of a predefined Symbol Grounding (SG) function mapping raw observations to high-level symbolic representations, or prior knowledge of the temporal task. In this work, we propose a fully learnable version of Neural Reward Machines (NRM), which can learn both the SG function and the automaton end-to-end, removing any reliance on prior knowledge. Our approach is therefore as easily applicable as classic deep RL (DRL) approaches, while being far more explainable, because of the finite and compact nature of automata. Furthermore, we show that by integrating Fully Learnable Reward Machines (FLNRM) with DRL, our method outperforms previous approaches based on Recurrent Neural Networks (RNNs).

Fully Learnable Neural Reward Machines / Dewidar, Hazem; Umili, Elena. - 4142:(2025), pp. 59-66. ( 7th International Workshop on Artificial Intelligence and Formal Verification, Logic, Automata, and Synthesis Bologna, Italy ).

Fully Learnable Neural Reward Machines

Hazem Dewidar
;
Elena Umili
2025

Abstract

Non-Markovian Reinforcement Learning (RL) tasks present significant challenges, as agents must reason over entire trajectories of state-action pairs to make optimal decisions. A common strategy to address this is through symbolic formalisms, such as Linear Temporal Logic (LTL) or automata, which provide a structured way to express temporally extended objectives. However, these approaches often rely on restrictive assumptions -- such as the availability of a predefined Symbol Grounding (SG) function mapping raw observations to high-level symbolic representations, or prior knowledge of the temporal task. In this work, we propose a fully learnable version of Neural Reward Machines (NRM), which can learn both the SG function and the automaton end-to-end, removing any reliance on prior knowledge. Our approach is therefore as easily applicable as classic deep RL (DRL) approaches, while being far more explainable, because of the finite and compact nature of automata. Furthermore, we show that by integrating Fully Learnable Reward Machines (FLNRM) with DRL, our method outperforms previous approaches based on Recurrent Neural Networks (RNNs).
2025
7th International Workshop on Artificial Intelligence and Formal Verification, Logic, Automata, and Synthesis
Automata Learning; Neurosymbolic learning; Deep Reinforcement Learning
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Fully Learnable Neural Reward Machines / Dewidar, Hazem; Umili, Elena. - 4142:(2025), pp. 59-66. ( 7th International Workshop on Artificial Intelligence and Formal Verification, Logic, Automata, and Synthesis Bologna, Italy ).
File allegati a questo prodotto
File Dimensione Formato  
Dewidar_Fully-Learnable-Neural_2025.pdf

accesso aperto

Note: https://ceur-ws.org/Vol-4142/
Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Creative commons
Dimensione 17.34 MB
Formato Adobe PDF
17.34 MB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1759804
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact