Quantum Recurrent Neural Networks are receiving an increased attention thanks to their enhanced generalization capabilities in time series analysis. However, their performances were bottlenecked by long training times and unscalable architectures. In this paper, we propose a novel Quantum Recurrent Neural Network model based on Quantum Gated Recurrent Units. It uses a learnable Variational Quantum Layer to process temporal data, interspersed with two classical layers to properly match the dimensionality of the input and output vectors. Such an architecture has fewer quantum parameters than existing Quantum Long Short-Term Memory models. Both the quantum networks were evaluated on periodic and real-world time series datasets, together with the classical counterparts. The quantum models exhibited superior performances compared to the classical ones in all the test cases. The Quantum Gated Recurrent Units outperformed the Quantum Long Short-Term Memory network despite having a simpler internal configuration. Moreover, the Quantum Gated Recurrent Units network demonstrated to be about 25% faster during the training and inference procedure over the Quantum Long Short-Term Memory. This improvement in speed comes with one less quantum circuit to be executed, suggesting that our model may offer a more efficient alternative for implementing Quantum Recurrent Neural Networks on both simulated and real quantum hardware.
A variational approach to quantum gated recurrent units / Ceschini, A.; Rosato, A.; Panella, M.. - In: JOURNAL OF PHYSICS COMMUNICATIONS. - ISSN 2399-6528. - 8:8(2024). [10.1088/2399-6528/ad6db7]
A variational approach to quantum gated recurrent units
Ceschini A.;Rosato A.;Panella M.
2024
Abstract
Quantum Recurrent Neural Networks are receiving an increased attention thanks to their enhanced generalization capabilities in time series analysis. However, their performances were bottlenecked by long training times and unscalable architectures. In this paper, we propose a novel Quantum Recurrent Neural Network model based on Quantum Gated Recurrent Units. It uses a learnable Variational Quantum Layer to process temporal data, interspersed with two classical layers to properly match the dimensionality of the input and output vectors. Such an architecture has fewer quantum parameters than existing Quantum Long Short-Term Memory models. Both the quantum networks were evaluated on periodic and real-world time series datasets, together with the classical counterparts. The quantum models exhibited superior performances compared to the classical ones in all the test cases. The Quantum Gated Recurrent Units outperformed the Quantum Long Short-Term Memory network despite having a simpler internal configuration. Moreover, the Quantum Gated Recurrent Units network demonstrated to be about 25% faster during the training and inference procedure over the Quantum Long Short-Term Memory. This improvement in speed comes with one less quantum circuit to be executed, suggesting that our model may offer a more efficient alternative for implementing Quantum Recurrent Neural Networks on both simulated and real quantum hardware.File | Dimensione | Formato | |
---|---|---|---|
Ceschini_Variational_2024.pdf
accesso aperto
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Creative commons
Dimensione
1.27 MB
Formato
Adobe PDF
|
1.27 MB | Adobe PDF |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.