Cargando…

Computing Generalized Matrix Inverse on Spiking Neural Substrate

Emerging neural hardware substrates, such as IBM's TrueNorth Neurosynaptic System, can provide an appealing platform for deploying numerical algorithms. For example, a recurrent Hopfield neural network can be used to find the Moore-Penrose generalized inverse of a matrix, thus enabling a broad...

Descripción completa

Detalles Bibliográficos
Autores principales: Shukla, Rohit, Khoram, Soroosh, Jorgensen, Erik, Li, Jing, Lipasti, Mikko, Wright, Stephen
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5859154/
https://www.ncbi.nlm.nih.gov/pubmed/29593483
http://dx.doi.org/10.3389/fnins.2018.00115
_version_ 1783307762594217984
author Shukla, Rohit
Khoram, Soroosh
Jorgensen, Erik
Li, Jing
Lipasti, Mikko
Wright, Stephen
author_facet Shukla, Rohit
Khoram, Soroosh
Jorgensen, Erik
Li, Jing
Lipasti, Mikko
Wright, Stephen
author_sort Shukla, Rohit
collection PubMed
description Emerging neural hardware substrates, such as IBM's TrueNorth Neurosynaptic System, can provide an appealing platform for deploying numerical algorithms. For example, a recurrent Hopfield neural network can be used to find the Moore-Penrose generalized inverse of a matrix, thus enabling a broad class of linear optimizations to be solved efficiently, at low energy cost. However, deploying numerical algorithms on hardware platforms that severely limit the range and precision of representation for numeric quantities can be quite challenging. This paper discusses these challenges and proposes a rigorous mathematical framework for reasoning about range and precision on such substrates. The paper derives techniques for normalizing inputs and properly quantizing synaptic weights originating from arbitrary systems of linear equations, so that solvers for those systems can be implemented in a provably correct manner on hardware-constrained neural substrates. The analytical model is empirically validated on the IBM TrueNorth platform, and results show that the guarantees provided by the framework for range and precision hold under experimental conditions. Experiments with optical flow demonstrate the energy benefits of deploying a reduced-precision and energy-efficient generalized matrix inverse engine on the IBM TrueNorth platform, reflecting 10× to 100× improvement over FPGA and ARM core baselines.
format Online
Article
Text
id pubmed-5859154
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-58591542018-03-28 Computing Generalized Matrix Inverse on Spiking Neural Substrate Shukla, Rohit Khoram, Soroosh Jorgensen, Erik Li, Jing Lipasti, Mikko Wright, Stephen Front Neurosci Neuroscience Emerging neural hardware substrates, such as IBM's TrueNorth Neurosynaptic System, can provide an appealing platform for deploying numerical algorithms. For example, a recurrent Hopfield neural network can be used to find the Moore-Penrose generalized inverse of a matrix, thus enabling a broad class of linear optimizations to be solved efficiently, at low energy cost. However, deploying numerical algorithms on hardware platforms that severely limit the range and precision of representation for numeric quantities can be quite challenging. This paper discusses these challenges and proposes a rigorous mathematical framework for reasoning about range and precision on such substrates. The paper derives techniques for normalizing inputs and properly quantizing synaptic weights originating from arbitrary systems of linear equations, so that solvers for those systems can be implemented in a provably correct manner on hardware-constrained neural substrates. The analytical model is empirically validated on the IBM TrueNorth platform, and results show that the guarantees provided by the framework for range and precision hold under experimental conditions. Experiments with optical flow demonstrate the energy benefits of deploying a reduced-precision and energy-efficient generalized matrix inverse engine on the IBM TrueNorth platform, reflecting 10× to 100× improvement over FPGA and ARM core baselines. Frontiers Media S.A. 2018-03-13 /pmc/articles/PMC5859154/ /pubmed/29593483 http://dx.doi.org/10.3389/fnins.2018.00115 Text en Copyright © 2018 Shukla, Khoram, Jorgensen, Li, Lipasti and Wright. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Shukla, Rohit
Khoram, Soroosh
Jorgensen, Erik
Li, Jing
Lipasti, Mikko
Wright, Stephen
Computing Generalized Matrix Inverse on Spiking Neural Substrate
title Computing Generalized Matrix Inverse on Spiking Neural Substrate
title_full Computing Generalized Matrix Inverse on Spiking Neural Substrate
title_fullStr Computing Generalized Matrix Inverse on Spiking Neural Substrate
title_full_unstemmed Computing Generalized Matrix Inverse on Spiking Neural Substrate
title_short Computing Generalized Matrix Inverse on Spiking Neural Substrate
title_sort computing generalized matrix inverse on spiking neural substrate
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5859154/
https://www.ncbi.nlm.nih.gov/pubmed/29593483
http://dx.doi.org/10.3389/fnins.2018.00115
work_keys_str_mv AT shuklarohit computinggeneralizedmatrixinverseonspikingneuralsubstrate
AT khoramsoroosh computinggeneralizedmatrixinverseonspikingneuralsubstrate
AT jorgensenerik computinggeneralizedmatrixinverseonspikingneuralsubstrate
AT lijing computinggeneralizedmatrixinverseonspikingneuralsubstrate
AT lipastimikko computinggeneralizedmatrixinverseonspikingneuralsubstrate
AT wrightstephen computinggeneralizedmatrixinverseonspikingneuralsubstrate