Event cameras are promising sensors for on-line and real-time vision tasks due to their high temporal resolution, low latency, and redundant static data elimination. Many vision algorithms use some form of spatial convolution (i.e. spatial pattern detection) as a fundamental component. However, additional consideration must be taken for event cameras, as the visual signal is asynchronous and sparse. While elegant methods have been proposed for event-based convolutions, they are unsuitable for real scenarios due to their inefficient processing pipeline and subsequent low event-throughput. This paper presents an efficient implementation based on decoupling the event-based computations from the computationally heavy convolutions, increasing the maximum event processing rate by 15. 92 × to over 10 million events/second, while still maintaining the event-based paradigm of asynchronous input and output. Results on public datasets with modern 640 × 480 event-camera recordings show that the proposed implementation achieves real-time processing with minimal impact on the convolution result, while the prior state-of-the-art results in a latency of over 1 second.
High-Throughput Asynchronous Convolutions for High-Resolution Event-Cameras / DE SOUZA ROSA, Leandro; Dinale, Aiko; Bamford, Simeon; Bartolozzi, Chiara; Glover, Arren. - (2022). (Intervento presentato al convegno 2022 8th International Conference on Event-Based Control, Communication, and Signal Processing (EBCCSP) tenutosi a Krakow, Poland) [10.1109/ebccsp56922.2022.9845500].
High-Throughput Asynchronous Convolutions for High-Resolution Event-Cameras
Leandro de Souza Rosa;
2022
Abstract
Event cameras are promising sensors for on-line and real-time vision tasks due to their high temporal resolution, low latency, and redundant static data elimination. Many vision algorithms use some form of spatial convolution (i.e. spatial pattern detection) as a fundamental component. However, additional consideration must be taken for event cameras, as the visual signal is asynchronous and sparse. While elegant methods have been proposed for event-based convolutions, they are unsuitable for real scenarios due to their inefficient processing pipeline and subsequent low event-throughput. This paper presents an efficient implementation based on decoupling the event-based computations from the computationally heavy convolutions, increasing the maximum event processing rate by 15. 92 × to over 10 million events/second, while still maintaining the event-based paradigm of asynchronous input and output. Results on public datasets with modern 640 × 480 event-camera recordings show that the proposed implementation achieves real-time processing with minimal impact on the convolution result, while the prior state-of-the-art results in a latency of over 1 second.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.