Cargando…

Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices

In a previous work we have detailed the requirements for obtaining maximal deep learning performance benefit by implementing fully connected deep neural networks (DNN) in the form of arrays of resistive devices. Here we extend the concept of Resistive Processing Unit (RPU) devices to convolutional n...

Descripción completa

Detalles Bibliográficos
Autores principales: Gokmen, Tayfun, Onen, Murat, Haensch, Wilfried
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5641314/
https://www.ncbi.nlm.nih.gov/pubmed/29066942
http://dx.doi.org/10.3389/fnins.2017.00538
_version_ 1783271198534139904
author Gokmen, Tayfun
Onen, Murat
Haensch, Wilfried
author_facet Gokmen, Tayfun
Onen, Murat
Haensch, Wilfried
author_sort Gokmen, Tayfun
collection PubMed
description In a previous work we have detailed the requirements for obtaining maximal deep learning performance benefit by implementing fully connected deep neural networks (DNN) in the form of arrays of resistive devices. Here we extend the concept of Resistive Processing Unit (RPU) devices to convolutional neural networks (CNNs). We show how to map the convolutional layers to fully connected RPU arrays such that the parallelism of the hardware can be fully utilized in all three cycles of the backpropagation algorithm. We find that the noise and bound limitations imposed by the analog nature of the computations performed on the arrays significantly affect the training accuracy of the CNNs. Noise and bound management techniques are presented that mitigate these problems without introducing any additional complexity in the analog circuits and that can be addressed by the digital circuits. In addition, we discuss digitally programmable update management and device variability reduction techniques that can be used selectively for some of the layers in a CNN. We show that a combination of all those techniques enables a successful application of the RPU concept for training CNNs. The techniques discussed here are more general and can be applied beyond CNN architectures and therefore enables applicability of the RPU approach to a large class of neural network architectures.
format Online
Article
Text
id pubmed-5641314
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-56413142017-10-24 Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices Gokmen, Tayfun Onen, Murat Haensch, Wilfried Front Neurosci Neuroscience In a previous work we have detailed the requirements for obtaining maximal deep learning performance benefit by implementing fully connected deep neural networks (DNN) in the form of arrays of resistive devices. Here we extend the concept of Resistive Processing Unit (RPU) devices to convolutional neural networks (CNNs). We show how to map the convolutional layers to fully connected RPU arrays such that the parallelism of the hardware can be fully utilized in all three cycles of the backpropagation algorithm. We find that the noise and bound limitations imposed by the analog nature of the computations performed on the arrays significantly affect the training accuracy of the CNNs. Noise and bound management techniques are presented that mitigate these problems without introducing any additional complexity in the analog circuits and that can be addressed by the digital circuits. In addition, we discuss digitally programmable update management and device variability reduction techniques that can be used selectively for some of the layers in a CNN. We show that a combination of all those techniques enables a successful application of the RPU concept for training CNNs. The techniques discussed here are more general and can be applied beyond CNN architectures and therefore enables applicability of the RPU approach to a large class of neural network architectures. Frontiers Media S.A. 2017-10-10 /pmc/articles/PMC5641314/ /pubmed/29066942 http://dx.doi.org/10.3389/fnins.2017.00538 Text en Copyright © 2017 Gokmen, Onen and Haensch. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Gokmen, Tayfun
Onen, Murat
Haensch, Wilfried
Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices
title Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices
title_full Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices
title_fullStr Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices
title_full_unstemmed Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices
title_short Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices
title_sort training deep convolutional neural networks with resistive cross-point devices
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5641314/
https://www.ncbi.nlm.nih.gov/pubmed/29066942
http://dx.doi.org/10.3389/fnins.2017.00538
work_keys_str_mv AT gokmentayfun trainingdeepconvolutionalneuralnetworkswithresistivecrosspointdevices
AT onenmurat trainingdeepconvolutionalneuralnetworkswithresistivecrosspointdevices
AT haenschwilfried trainingdeepconvolutionalneuralnetworkswithresistivecrosspointdevices