Cargando…
Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices
In a previous work we have detailed the requirements for obtaining maximal deep learning performance benefit by implementing fully connected deep neural networks (DNN) in the form of arrays of resistive devices. Here we extend the concept of Resistive Processing Unit (RPU) devices to convolutional n...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5641314/ https://www.ncbi.nlm.nih.gov/pubmed/29066942 http://dx.doi.org/10.3389/fnins.2017.00538 |
_version_ | 1783271198534139904 |
---|---|
author | Gokmen, Tayfun Onen, Murat Haensch, Wilfried |
author_facet | Gokmen, Tayfun Onen, Murat Haensch, Wilfried |
author_sort | Gokmen, Tayfun |
collection | PubMed |
description | In a previous work we have detailed the requirements for obtaining maximal deep learning performance benefit by implementing fully connected deep neural networks (DNN) in the form of arrays of resistive devices. Here we extend the concept of Resistive Processing Unit (RPU) devices to convolutional neural networks (CNNs). We show how to map the convolutional layers to fully connected RPU arrays such that the parallelism of the hardware can be fully utilized in all three cycles of the backpropagation algorithm. We find that the noise and bound limitations imposed by the analog nature of the computations performed on the arrays significantly affect the training accuracy of the CNNs. Noise and bound management techniques are presented that mitigate these problems without introducing any additional complexity in the analog circuits and that can be addressed by the digital circuits. In addition, we discuss digitally programmable update management and device variability reduction techniques that can be used selectively for some of the layers in a CNN. We show that a combination of all those techniques enables a successful application of the RPU concept for training CNNs. The techniques discussed here are more general and can be applied beyond CNN architectures and therefore enables applicability of the RPU approach to a large class of neural network architectures. |
format | Online Article Text |
id | pubmed-5641314 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-56413142017-10-24 Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices Gokmen, Tayfun Onen, Murat Haensch, Wilfried Front Neurosci Neuroscience In a previous work we have detailed the requirements for obtaining maximal deep learning performance benefit by implementing fully connected deep neural networks (DNN) in the form of arrays of resistive devices. Here we extend the concept of Resistive Processing Unit (RPU) devices to convolutional neural networks (CNNs). We show how to map the convolutional layers to fully connected RPU arrays such that the parallelism of the hardware can be fully utilized in all three cycles of the backpropagation algorithm. We find that the noise and bound limitations imposed by the analog nature of the computations performed on the arrays significantly affect the training accuracy of the CNNs. Noise and bound management techniques are presented that mitigate these problems without introducing any additional complexity in the analog circuits and that can be addressed by the digital circuits. In addition, we discuss digitally programmable update management and device variability reduction techniques that can be used selectively for some of the layers in a CNN. We show that a combination of all those techniques enables a successful application of the RPU concept for training CNNs. The techniques discussed here are more general and can be applied beyond CNN architectures and therefore enables applicability of the RPU approach to a large class of neural network architectures. Frontiers Media S.A. 2017-10-10 /pmc/articles/PMC5641314/ /pubmed/29066942 http://dx.doi.org/10.3389/fnins.2017.00538 Text en Copyright © 2017 Gokmen, Onen and Haensch. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Gokmen, Tayfun Onen, Murat Haensch, Wilfried Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices |
title | Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices |
title_full | Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices |
title_fullStr | Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices |
title_full_unstemmed | Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices |
title_short | Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices |
title_sort | training deep convolutional neural networks with resistive cross-point devices |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5641314/ https://www.ncbi.nlm.nih.gov/pubmed/29066942 http://dx.doi.org/10.3389/fnins.2017.00538 |
work_keys_str_mv | AT gokmentayfun trainingdeepconvolutionalneuralnetworkswithresistivecrosspointdevices AT onenmurat trainingdeepconvolutionalneuralnetworkswithresistivecrosspointdevices AT haenschwilfried trainingdeepconvolutionalneuralnetworkswithresistivecrosspointdevices |