Cargando…

Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware

Spatio-temporal pattern recognition is a fundamental ability of the brain which is required for numerous real-world activities. Recent deep learning approaches have reached outstanding accuracies in such tasks, but their implementation on conventional embedded solutions is still very computationally...

Descripción completa

Detalles Bibliográficos
Autores principales: Müller-Cleve, Simon F., Fra, Vittorio, Khacef, Lyes, Pequeño-Zurro, Alejandro, Klepatsch, Daniel, Forno, Evelina, Ivanovich, Diego G., Rastogi, Shavika, Urgese, Gianvito, Zenke, Friedemann, Bartolozzi, Chiara
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9695069/
https://www.ncbi.nlm.nih.gov/pubmed/36440280
http://dx.doi.org/10.3389/fnins.2022.951164
_version_ 1784837964953550848
author Müller-Cleve, Simon F.
Fra, Vittorio
Khacef, Lyes
Pequeño-Zurro, Alejandro
Klepatsch, Daniel
Forno, Evelina
Ivanovich, Diego G.
Rastogi, Shavika
Urgese, Gianvito
Zenke, Friedemann
Bartolozzi, Chiara
author_facet Müller-Cleve, Simon F.
Fra, Vittorio
Khacef, Lyes
Pequeño-Zurro, Alejandro
Klepatsch, Daniel
Forno, Evelina
Ivanovich, Diego G.
Rastogi, Shavika
Urgese, Gianvito
Zenke, Friedemann
Bartolozzi, Chiara
author_sort Müller-Cleve, Simon F.
collection PubMed
description Spatio-temporal pattern recognition is a fundamental ability of the brain which is required for numerous real-world activities. Recent deep learning approaches have reached outstanding accuracies in such tasks, but their implementation on conventional embedded solutions is still very computationally and energy expensive. Tactile sensing in robotic applications is a representative example where real-time processing and energy efficiency are required. Following a brain-inspired computing approach, we propose a new benchmark for spatio-temporal tactile pattern recognition at the edge through Braille letter reading. We recorded a new Braille letters dataset based on the capacitive tactile sensors of the iCub robot's fingertip. We then investigated the importance of spatial and temporal information as well as the impact of event-based encoding on spike-based computation. Afterward, we trained and compared feedforward and recurrent Spiking Neural Networks (SNNs) offline using Backpropagation Through Time (BPTT) with surrogate gradients, then we deployed them on the Intel Loihi neuromorphic chip for fast and efficient inference. We compared our approach to standard classifiers, in particular to the Long Short-Term Memory (LSTM) deployed on the embedded NVIDIA Jetson GPU, in terms of classification accuracy, power, and energy consumption together with computational delay. Our results show that the LSTM reaches ~97% of accuracy, outperforming the recurrent SNN by ~17% when using continuous frame-based data instead of event-based inputs. However, the recurrent SNN on Loihi with event-based inputs is ~500 times more energy-efficient than the LSTM on Jetson, requiring a total power of only ~30 mW. This work proposes a new benchmark for tactile sensing and highlights the challenges and opportunities of event-based encoding, neuromorphic hardware, and spike-based computing for spatio-temporal pattern recognition at the edge.
format Online
Article
Text
id pubmed-9695069
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-96950692022-11-26 Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware Müller-Cleve, Simon F. Fra, Vittorio Khacef, Lyes Pequeño-Zurro, Alejandro Klepatsch, Daniel Forno, Evelina Ivanovich, Diego G. Rastogi, Shavika Urgese, Gianvito Zenke, Friedemann Bartolozzi, Chiara Front Neurosci Neuroscience Spatio-temporal pattern recognition is a fundamental ability of the brain which is required for numerous real-world activities. Recent deep learning approaches have reached outstanding accuracies in such tasks, but their implementation on conventional embedded solutions is still very computationally and energy expensive. Tactile sensing in robotic applications is a representative example where real-time processing and energy efficiency are required. Following a brain-inspired computing approach, we propose a new benchmark for spatio-temporal tactile pattern recognition at the edge through Braille letter reading. We recorded a new Braille letters dataset based on the capacitive tactile sensors of the iCub robot's fingertip. We then investigated the importance of spatial and temporal information as well as the impact of event-based encoding on spike-based computation. Afterward, we trained and compared feedforward and recurrent Spiking Neural Networks (SNNs) offline using Backpropagation Through Time (BPTT) with surrogate gradients, then we deployed them on the Intel Loihi neuromorphic chip for fast and efficient inference. We compared our approach to standard classifiers, in particular to the Long Short-Term Memory (LSTM) deployed on the embedded NVIDIA Jetson GPU, in terms of classification accuracy, power, and energy consumption together with computational delay. Our results show that the LSTM reaches ~97% of accuracy, outperforming the recurrent SNN by ~17% when using continuous frame-based data instead of event-based inputs. However, the recurrent SNN on Loihi with event-based inputs is ~500 times more energy-efficient than the LSTM on Jetson, requiring a total power of only ~30 mW. This work proposes a new benchmark for tactile sensing and highlights the challenges and opportunities of event-based encoding, neuromorphic hardware, and spike-based computing for spatio-temporal pattern recognition at the edge. Frontiers Media S.A. 2022-11-11 /pmc/articles/PMC9695069/ /pubmed/36440280 http://dx.doi.org/10.3389/fnins.2022.951164 Text en Copyright © 2022 Müller-Cleve, Fra, Khacef, Pequeño-Zurro, Klepatsch, Forno, Ivanovich, Rastogi, Urgese, Zenke and Bartolozzi. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Müller-Cleve, Simon F.
Fra, Vittorio
Khacef, Lyes
Pequeño-Zurro, Alejandro
Klepatsch, Daniel
Forno, Evelina
Ivanovich, Diego G.
Rastogi, Shavika
Urgese, Gianvito
Zenke, Friedemann
Bartolozzi, Chiara
Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware
title Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware
title_full Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware
title_fullStr Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware
title_full_unstemmed Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware
title_short Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware
title_sort braille letter reading: a benchmark for spatio-temporal pattern recognition on neuromorphic hardware
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9695069/
https://www.ncbi.nlm.nih.gov/pubmed/36440280
http://dx.doi.org/10.3389/fnins.2022.951164
work_keys_str_mv AT mullerclevesimonf brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT fravittorio brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT khaceflyes brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT pequenozurroalejandro brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT klepatschdaniel brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT fornoevelina brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT ivanovichdiegog brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT rastogishavika brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT urgesegianvito brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT zenkefriedemann brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware
AT bartolozzichiara brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware