Cargando…

Direct Feedback Alignment With Sparse Connections for Local Learning

Recent advances in deep neural networks (DNNs) owe their success to training algorithms that use backpropagation and gradient-descent. Backpropagation, while highly effective on von Neumann architectures, becomes inefficient when scaling to large networks. Commonly referred to as the weight transpor...

Descripción completa

Detalles Bibliográficos
Autores principales: Crafton, Brian, Parihar, Abhinav, Gebhardt, Evan, Raychowdhury, Arijit
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6542988/
https://www.ncbi.nlm.nih.gov/pubmed/31178689
http://dx.doi.org/10.3389/fnins.2019.00525
_version_ 1783423011308699648
author Crafton, Brian
Parihar, Abhinav
Gebhardt, Evan
Raychowdhury, Arijit
author_facet Crafton, Brian
Parihar, Abhinav
Gebhardt, Evan
Raychowdhury, Arijit
author_sort Crafton, Brian
collection PubMed
description Recent advances in deep neural networks (DNNs) owe their success to training algorithms that use backpropagation and gradient-descent. Backpropagation, while highly effective on von Neumann architectures, becomes inefficient when scaling to large networks. Commonly referred to as the weight transport problem, each neuron's dependence on the weights and errors located deeper in the network require exhaustive data movement which presents a key problem in enhancing the performance and energy-efficiency of machine-learning hardware. In this work, we propose a bio-plausible alternative to backpropagation drawing from advances in feedback alignment algorithms in which the error computation at a single synapse reduces to the product of three scalar values. Using a sparse feedback matrix, we show that a neuron needs only a fraction of the information previously used by the feedback alignment algorithms. Consequently, memory and compute can be partitioned and distributed whichever way produces the most efficient forward pass so long as a single error can be delivered to each neuron. We evaluate our algorithm using standard datasets, including ImageNet, to address the concern of scaling to challenging problems. Our results show orders of magnitude improvement in data movement and 2× improvement in multiply-and-accumulate operations over backpropagation. Like previous work, we observe that any variant of feedback alignment suffers significant losses in classification accuracy on deep convolutional neural networks. By transferring trained convolutional layers and training the fully connected layers using direct feedback alignment, we demonstrate that direct feedback alignment can obtain results competitive with backpropagation. Furthermore, we observe that using an extremely sparse feedback matrix, rather than a dense one, results in a small accuracy drop while yielding hardware advantages. All the code and results are available under https://github.com/bcrafton/ssdfa.
format Online
Article
Text
id pubmed-6542988
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-65429882019-06-07 Direct Feedback Alignment With Sparse Connections for Local Learning Crafton, Brian Parihar, Abhinav Gebhardt, Evan Raychowdhury, Arijit Front Neurosci Neuroscience Recent advances in deep neural networks (DNNs) owe their success to training algorithms that use backpropagation and gradient-descent. Backpropagation, while highly effective on von Neumann architectures, becomes inefficient when scaling to large networks. Commonly referred to as the weight transport problem, each neuron's dependence on the weights and errors located deeper in the network require exhaustive data movement which presents a key problem in enhancing the performance and energy-efficiency of machine-learning hardware. In this work, we propose a bio-plausible alternative to backpropagation drawing from advances in feedback alignment algorithms in which the error computation at a single synapse reduces to the product of three scalar values. Using a sparse feedback matrix, we show that a neuron needs only a fraction of the information previously used by the feedback alignment algorithms. Consequently, memory and compute can be partitioned and distributed whichever way produces the most efficient forward pass so long as a single error can be delivered to each neuron. We evaluate our algorithm using standard datasets, including ImageNet, to address the concern of scaling to challenging problems. Our results show orders of magnitude improvement in data movement and 2× improvement in multiply-and-accumulate operations over backpropagation. Like previous work, we observe that any variant of feedback alignment suffers significant losses in classification accuracy on deep convolutional neural networks. By transferring trained convolutional layers and training the fully connected layers using direct feedback alignment, we demonstrate that direct feedback alignment can obtain results competitive with backpropagation. Furthermore, we observe that using an extremely sparse feedback matrix, rather than a dense one, results in a small accuracy drop while yielding hardware advantages. All the code and results are available under https://github.com/bcrafton/ssdfa. Frontiers Media S.A. 2019-05-24 /pmc/articles/PMC6542988/ /pubmed/31178689 http://dx.doi.org/10.3389/fnins.2019.00525 Text en Copyright © 2019 Crafton, Parihar, Gebhardt and Raychowdhury. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Crafton, Brian
Parihar, Abhinav
Gebhardt, Evan
Raychowdhury, Arijit
Direct Feedback Alignment With Sparse Connections for Local Learning
title Direct Feedback Alignment With Sparse Connections for Local Learning
title_full Direct Feedback Alignment With Sparse Connections for Local Learning
title_fullStr Direct Feedback Alignment With Sparse Connections for Local Learning
title_full_unstemmed Direct Feedback Alignment With Sparse Connections for Local Learning
title_short Direct Feedback Alignment With Sparse Connections for Local Learning
title_sort direct feedback alignment with sparse connections for local learning
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6542988/
https://www.ncbi.nlm.nih.gov/pubmed/31178689
http://dx.doi.org/10.3389/fnins.2019.00525
work_keys_str_mv AT craftonbrian directfeedbackalignmentwithsparseconnectionsforlocallearning
AT pariharabhinav directfeedbackalignmentwithsparseconnectionsforlocallearning
AT gebhardtevan directfeedbackalignmentwithsparseconnectionsforlocallearning
AT raychowdhuryarijit directfeedbackalignmentwithsparseconnectionsforlocallearning