Cargando…

Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks

Artificial neural networks (ANNs) trained using backpropagation are powerful learning architectures that have achieved state-of-the-art performance in various benchmarks. Significant effort has been devoted to developing custom silicon devices to accelerate inference in ANNs. Accelerating the traini...

Descripción completa

Detalles Bibliográficos
Autores principales: Mostafa, Hesham, Pedroni, Bruno, Sheik, Sadique, Cauwenberghs, Gert
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5592276/
https://www.ncbi.nlm.nih.gov/pubmed/28932180
http://dx.doi.org/10.3389/fnins.2017.00496
_version_ 1783262858394468352
author Mostafa, Hesham
Pedroni, Bruno
Sheik, Sadique
Cauwenberghs, Gert
author_facet Mostafa, Hesham
Pedroni, Bruno
Sheik, Sadique
Cauwenberghs, Gert
author_sort Mostafa, Hesham
collection PubMed
description Artificial neural networks (ANNs) trained using backpropagation are powerful learning architectures that have achieved state-of-the-art performance in various benchmarks. Significant effort has been devoted to developing custom silicon devices to accelerate inference in ANNs. Accelerating the training phase, however, has attracted relatively little attention. In this paper, we describe a hardware-efficient on-line learning technique for feedforward multi-layer ANNs that is based on pipelined backpropagation. Learning is performed in parallel with inference in the forward pass, removing the need for an explicit backward pass and requiring no extra weight lookup. By using binary state variables in the feedforward network and ternary errors in truncated-error backpropagation, the need for any multiplications in the forward and backward passes is removed, and memory requirements for the pipelining are drastically reduced. Further reduction in addition operations owing to the sparsity in the forward neural and backpropagating error signal paths contributes to highly efficient hardware implementation. For proof-of-concept validation, we demonstrate on-line learning of MNIST handwritten digit classification on a Spartan 6 FPGA interfacing with an external 1Gb DDR2 DRAM, that shows small degradation in test error performance compared to an equivalently sized binary ANN trained off-line using standard back-propagation and exact errors. Our results highlight an attractive synergy between pipelined backpropagation and binary-state networks in substantially reducing computation and memory requirements, making pipelined on-line learning practical in deep networks.
format Online
Article
Text
id pubmed-5592276
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-55922762017-09-20 Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks Mostafa, Hesham Pedroni, Bruno Sheik, Sadique Cauwenberghs, Gert Front Neurosci Neuroscience Artificial neural networks (ANNs) trained using backpropagation are powerful learning architectures that have achieved state-of-the-art performance in various benchmarks. Significant effort has been devoted to developing custom silicon devices to accelerate inference in ANNs. Accelerating the training phase, however, has attracted relatively little attention. In this paper, we describe a hardware-efficient on-line learning technique for feedforward multi-layer ANNs that is based on pipelined backpropagation. Learning is performed in parallel with inference in the forward pass, removing the need for an explicit backward pass and requiring no extra weight lookup. By using binary state variables in the feedforward network and ternary errors in truncated-error backpropagation, the need for any multiplications in the forward and backward passes is removed, and memory requirements for the pipelining are drastically reduced. Further reduction in addition operations owing to the sparsity in the forward neural and backpropagating error signal paths contributes to highly efficient hardware implementation. For proof-of-concept validation, we demonstrate on-line learning of MNIST handwritten digit classification on a Spartan 6 FPGA interfacing with an external 1Gb DDR2 DRAM, that shows small degradation in test error performance compared to an equivalently sized binary ANN trained off-line using standard back-propagation and exact errors. Our results highlight an attractive synergy between pipelined backpropagation and binary-state networks in substantially reducing computation and memory requirements, making pipelined on-line learning practical in deep networks. Frontiers Media S.A. 2017-09-06 /pmc/articles/PMC5592276/ /pubmed/28932180 http://dx.doi.org/10.3389/fnins.2017.00496 Text en Copyright © 2017 Mostafa, Pedroni, Sheik and Cauwenberghs. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Mostafa, Hesham
Pedroni, Bruno
Sheik, Sadique
Cauwenberghs, Gert
Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks
title Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks
title_full Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks
title_fullStr Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks
title_full_unstemmed Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks
title_short Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks
title_sort hardware-efficient on-line learning through pipelined truncated-error backpropagation in binary-state networks
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5592276/
https://www.ncbi.nlm.nih.gov/pubmed/28932180
http://dx.doi.org/10.3389/fnins.2017.00496
work_keys_str_mv AT mostafahesham hardwareefficientonlinelearningthroughpipelinedtruncatederrorbackpropagationinbinarystatenetworks
AT pedronibruno hardwareefficientonlinelearningthroughpipelinedtruncatederrorbackpropagationinbinarystatenetworks
AT sheiksadique hardwareefficientonlinelearningthroughpipelinedtruncatederrorbackpropagationinbinarystatenetworks
AT cauwenberghsgert hardwareefficientonlinelearningthroughpipelinedtruncatederrorbackpropagationinbinarystatenetworks