Cargando…
Direct Feedback Alignment With Sparse Connections for Local Learning
Recent advances in deep neural networks (DNNs) owe their success to training algorithms that use backpropagation and gradient-descent. Backpropagation, while highly effective on von Neumann architectures, becomes inefficient when scaling to large networks. Commonly referred to as the weight transpor...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6542988/ https://www.ncbi.nlm.nih.gov/pubmed/31178689 http://dx.doi.org/10.3389/fnins.2019.00525 |
_version_ | 1783423011308699648 |
---|---|
author | Crafton, Brian Parihar, Abhinav Gebhardt, Evan Raychowdhury, Arijit |
author_facet | Crafton, Brian Parihar, Abhinav Gebhardt, Evan Raychowdhury, Arijit |
author_sort | Crafton, Brian |
collection | PubMed |
description | Recent advances in deep neural networks (DNNs) owe their success to training algorithms that use backpropagation and gradient-descent. Backpropagation, while highly effective on von Neumann architectures, becomes inefficient when scaling to large networks. Commonly referred to as the weight transport problem, each neuron's dependence on the weights and errors located deeper in the network require exhaustive data movement which presents a key problem in enhancing the performance and energy-efficiency of machine-learning hardware. In this work, we propose a bio-plausible alternative to backpropagation drawing from advances in feedback alignment algorithms in which the error computation at a single synapse reduces to the product of three scalar values. Using a sparse feedback matrix, we show that a neuron needs only a fraction of the information previously used by the feedback alignment algorithms. Consequently, memory and compute can be partitioned and distributed whichever way produces the most efficient forward pass so long as a single error can be delivered to each neuron. We evaluate our algorithm using standard datasets, including ImageNet, to address the concern of scaling to challenging problems. Our results show orders of magnitude improvement in data movement and 2× improvement in multiply-and-accumulate operations over backpropagation. Like previous work, we observe that any variant of feedback alignment suffers significant losses in classification accuracy on deep convolutional neural networks. By transferring trained convolutional layers and training the fully connected layers using direct feedback alignment, we demonstrate that direct feedback alignment can obtain results competitive with backpropagation. Furthermore, we observe that using an extremely sparse feedback matrix, rather than a dense one, results in a small accuracy drop while yielding hardware advantages. All the code and results are available under https://github.com/bcrafton/ssdfa. |
format | Online Article Text |
id | pubmed-6542988 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-65429882019-06-07 Direct Feedback Alignment With Sparse Connections for Local Learning Crafton, Brian Parihar, Abhinav Gebhardt, Evan Raychowdhury, Arijit Front Neurosci Neuroscience Recent advances in deep neural networks (DNNs) owe their success to training algorithms that use backpropagation and gradient-descent. Backpropagation, while highly effective on von Neumann architectures, becomes inefficient when scaling to large networks. Commonly referred to as the weight transport problem, each neuron's dependence on the weights and errors located deeper in the network require exhaustive data movement which presents a key problem in enhancing the performance and energy-efficiency of machine-learning hardware. In this work, we propose a bio-plausible alternative to backpropagation drawing from advances in feedback alignment algorithms in which the error computation at a single synapse reduces to the product of three scalar values. Using a sparse feedback matrix, we show that a neuron needs only a fraction of the information previously used by the feedback alignment algorithms. Consequently, memory and compute can be partitioned and distributed whichever way produces the most efficient forward pass so long as a single error can be delivered to each neuron. We evaluate our algorithm using standard datasets, including ImageNet, to address the concern of scaling to challenging problems. Our results show orders of magnitude improvement in data movement and 2× improvement in multiply-and-accumulate operations over backpropagation. Like previous work, we observe that any variant of feedback alignment suffers significant losses in classification accuracy on deep convolutional neural networks. By transferring trained convolutional layers and training the fully connected layers using direct feedback alignment, we demonstrate that direct feedback alignment can obtain results competitive with backpropagation. Furthermore, we observe that using an extremely sparse feedback matrix, rather than a dense one, results in a small accuracy drop while yielding hardware advantages. All the code and results are available under https://github.com/bcrafton/ssdfa. Frontiers Media S.A. 2019-05-24 /pmc/articles/PMC6542988/ /pubmed/31178689 http://dx.doi.org/10.3389/fnins.2019.00525 Text en Copyright © 2019 Crafton, Parihar, Gebhardt and Raychowdhury. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Crafton, Brian Parihar, Abhinav Gebhardt, Evan Raychowdhury, Arijit Direct Feedback Alignment With Sparse Connections for Local Learning |
title | Direct Feedback Alignment With Sparse Connections for Local Learning |
title_full | Direct Feedback Alignment With Sparse Connections for Local Learning |
title_fullStr | Direct Feedback Alignment With Sparse Connections for Local Learning |
title_full_unstemmed | Direct Feedback Alignment With Sparse Connections for Local Learning |
title_short | Direct Feedback Alignment With Sparse Connections for Local Learning |
title_sort | direct feedback alignment with sparse connections for local learning |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6542988/ https://www.ncbi.nlm.nih.gov/pubmed/31178689 http://dx.doi.org/10.3389/fnins.2019.00525 |
work_keys_str_mv | AT craftonbrian directfeedbackalignmentwithsparseconnectionsforlocallearning AT pariharabhinav directfeedbackalignmentwithsparseconnectionsforlocallearning AT gebhardtevan directfeedbackalignmentwithsparseconnectionsforlocallearning AT raychowdhuryarijit directfeedbackalignmentwithsparseconnectionsforlocallearning |