Cargando…
Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware
Spatio-temporal pattern recognition is a fundamental ability of the brain which is required for numerous real-world activities. Recent deep learning approaches have reached outstanding accuracies in such tasks, but their implementation on conventional embedded solutions is still very computationally...
Autores principales: | , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9695069/ https://www.ncbi.nlm.nih.gov/pubmed/36440280 http://dx.doi.org/10.3389/fnins.2022.951164 |
_version_ | 1784837964953550848 |
---|---|
author | Müller-Cleve, Simon F. Fra, Vittorio Khacef, Lyes Pequeño-Zurro, Alejandro Klepatsch, Daniel Forno, Evelina Ivanovich, Diego G. Rastogi, Shavika Urgese, Gianvito Zenke, Friedemann Bartolozzi, Chiara |
author_facet | Müller-Cleve, Simon F. Fra, Vittorio Khacef, Lyes Pequeño-Zurro, Alejandro Klepatsch, Daniel Forno, Evelina Ivanovich, Diego G. Rastogi, Shavika Urgese, Gianvito Zenke, Friedemann Bartolozzi, Chiara |
author_sort | Müller-Cleve, Simon F. |
collection | PubMed |
description | Spatio-temporal pattern recognition is a fundamental ability of the brain which is required for numerous real-world activities. Recent deep learning approaches have reached outstanding accuracies in such tasks, but their implementation on conventional embedded solutions is still very computationally and energy expensive. Tactile sensing in robotic applications is a representative example where real-time processing and energy efficiency are required. Following a brain-inspired computing approach, we propose a new benchmark for spatio-temporal tactile pattern recognition at the edge through Braille letter reading. We recorded a new Braille letters dataset based on the capacitive tactile sensors of the iCub robot's fingertip. We then investigated the importance of spatial and temporal information as well as the impact of event-based encoding on spike-based computation. Afterward, we trained and compared feedforward and recurrent Spiking Neural Networks (SNNs) offline using Backpropagation Through Time (BPTT) with surrogate gradients, then we deployed them on the Intel Loihi neuromorphic chip for fast and efficient inference. We compared our approach to standard classifiers, in particular to the Long Short-Term Memory (LSTM) deployed on the embedded NVIDIA Jetson GPU, in terms of classification accuracy, power, and energy consumption together with computational delay. Our results show that the LSTM reaches ~97% of accuracy, outperforming the recurrent SNN by ~17% when using continuous frame-based data instead of event-based inputs. However, the recurrent SNN on Loihi with event-based inputs is ~500 times more energy-efficient than the LSTM on Jetson, requiring a total power of only ~30 mW. This work proposes a new benchmark for tactile sensing and highlights the challenges and opportunities of event-based encoding, neuromorphic hardware, and spike-based computing for spatio-temporal pattern recognition at the edge. |
format | Online Article Text |
id | pubmed-9695069 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-96950692022-11-26 Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware Müller-Cleve, Simon F. Fra, Vittorio Khacef, Lyes Pequeño-Zurro, Alejandro Klepatsch, Daniel Forno, Evelina Ivanovich, Diego G. Rastogi, Shavika Urgese, Gianvito Zenke, Friedemann Bartolozzi, Chiara Front Neurosci Neuroscience Spatio-temporal pattern recognition is a fundamental ability of the brain which is required for numerous real-world activities. Recent deep learning approaches have reached outstanding accuracies in such tasks, but their implementation on conventional embedded solutions is still very computationally and energy expensive. Tactile sensing in robotic applications is a representative example where real-time processing and energy efficiency are required. Following a brain-inspired computing approach, we propose a new benchmark for spatio-temporal tactile pattern recognition at the edge through Braille letter reading. We recorded a new Braille letters dataset based on the capacitive tactile sensors of the iCub robot's fingertip. We then investigated the importance of spatial and temporal information as well as the impact of event-based encoding on spike-based computation. Afterward, we trained and compared feedforward and recurrent Spiking Neural Networks (SNNs) offline using Backpropagation Through Time (BPTT) with surrogate gradients, then we deployed them on the Intel Loihi neuromorphic chip for fast and efficient inference. We compared our approach to standard classifiers, in particular to the Long Short-Term Memory (LSTM) deployed on the embedded NVIDIA Jetson GPU, in terms of classification accuracy, power, and energy consumption together with computational delay. Our results show that the LSTM reaches ~97% of accuracy, outperforming the recurrent SNN by ~17% when using continuous frame-based data instead of event-based inputs. However, the recurrent SNN on Loihi with event-based inputs is ~500 times more energy-efficient than the LSTM on Jetson, requiring a total power of only ~30 mW. This work proposes a new benchmark for tactile sensing and highlights the challenges and opportunities of event-based encoding, neuromorphic hardware, and spike-based computing for spatio-temporal pattern recognition at the edge. Frontiers Media S.A. 2022-11-11 /pmc/articles/PMC9695069/ /pubmed/36440280 http://dx.doi.org/10.3389/fnins.2022.951164 Text en Copyright © 2022 Müller-Cleve, Fra, Khacef, Pequeño-Zurro, Klepatsch, Forno, Ivanovich, Rastogi, Urgese, Zenke and Bartolozzi. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Müller-Cleve, Simon F. Fra, Vittorio Khacef, Lyes Pequeño-Zurro, Alejandro Klepatsch, Daniel Forno, Evelina Ivanovich, Diego G. Rastogi, Shavika Urgese, Gianvito Zenke, Friedemann Bartolozzi, Chiara Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware |
title | Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware |
title_full | Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware |
title_fullStr | Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware |
title_full_unstemmed | Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware |
title_short | Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware |
title_sort | braille letter reading: a benchmark for spatio-temporal pattern recognition on neuromorphic hardware |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9695069/ https://www.ncbi.nlm.nih.gov/pubmed/36440280 http://dx.doi.org/10.3389/fnins.2022.951164 |
work_keys_str_mv | AT mullerclevesimonf brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT fravittorio brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT khaceflyes brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT pequenozurroalejandro brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT klepatschdaniel brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT fornoevelina brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT ivanovichdiegog brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT rastogishavika brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT urgesegianvito brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT zenkefriedemann brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware AT bartolozzichiara brailleletterreadingabenchmarkforspatiotemporalpatternrecognitiononneuromorphichardware |