Cargando…

Neural Network Training With Asymmetric Crosspoint Elements

Analog crossbar arrays comprising programmable non-volatile resistors are under intense investigation for acceleration of deep neural network training. However, the ubiquitous asymmetric conductance modulation of practical resistive devices critically degrades the classification performance of netwo...

Descripción completa

Detalles Bibliográficos
Autores principales: Onen, Murat, Gokmen, Tayfun, Todorov, Teodor K., Nowicki, Tomasz, del Alamo, Jesús A., Rozen, John, Haensch, Wilfried, Kim, Seyoung
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9124763/
https://www.ncbi.nlm.nih.gov/pubmed/35615470
http://dx.doi.org/10.3389/frai.2022.891624
_version_ 1784711794624823296
author Onen, Murat
Gokmen, Tayfun
Todorov, Teodor K.
Nowicki, Tomasz
del Alamo, Jesús A.
Rozen, John
Haensch, Wilfried
Kim, Seyoung
author_facet Onen, Murat
Gokmen, Tayfun
Todorov, Teodor K.
Nowicki, Tomasz
del Alamo, Jesús A.
Rozen, John
Haensch, Wilfried
Kim, Seyoung
author_sort Onen, Murat
collection PubMed
description Analog crossbar arrays comprising programmable non-volatile resistors are under intense investigation for acceleration of deep neural network training. However, the ubiquitous asymmetric conductance modulation of practical resistive devices critically degrades the classification performance of networks trained with conventional algorithms. Here we first describe the fundamental reasons behind this incompatibility. Then, we explain the theoretical underpinnings of a novel fully-parallel training algorithm that is compatible with asymmetric crosspoint elements. By establishing a powerful analogy with classical mechanics, we explain how device asymmetry can be exploited as a useful feature for analog deep learning processors. Instead of conventionally tuning weights in the direction of the error function gradient, network parameters can be programmed to successfully minimize the total energy (Hamiltonian) of the system that incorporates the effects of device asymmetry. Our technique enables immediate realization of analog deep learning accelerators based on readily available device technologies.
format Online
Article
Text
id pubmed-9124763
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-91247632022-05-24 Neural Network Training With Asymmetric Crosspoint Elements Onen, Murat Gokmen, Tayfun Todorov, Teodor K. Nowicki, Tomasz del Alamo, Jesús A. Rozen, John Haensch, Wilfried Kim, Seyoung Front Artif Intell Artificial Intelligence Analog crossbar arrays comprising programmable non-volatile resistors are under intense investigation for acceleration of deep neural network training. However, the ubiquitous asymmetric conductance modulation of practical resistive devices critically degrades the classification performance of networks trained with conventional algorithms. Here we first describe the fundamental reasons behind this incompatibility. Then, we explain the theoretical underpinnings of a novel fully-parallel training algorithm that is compatible with asymmetric crosspoint elements. By establishing a powerful analogy with classical mechanics, we explain how device asymmetry can be exploited as a useful feature for analog deep learning processors. Instead of conventionally tuning weights in the direction of the error function gradient, network parameters can be programmed to successfully minimize the total energy (Hamiltonian) of the system that incorporates the effects of device asymmetry. Our technique enables immediate realization of analog deep learning accelerators based on readily available device technologies. Frontiers Media S.A. 2022-05-09 /pmc/articles/PMC9124763/ /pubmed/35615470 http://dx.doi.org/10.3389/frai.2022.891624 Text en Copyright © 2022 Onen, Gokmen, Todorov, Nowicki, del Alamo, Rozen, Haensch and Kim. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Onen, Murat
Gokmen, Tayfun
Todorov, Teodor K.
Nowicki, Tomasz
del Alamo, Jesús A.
Rozen, John
Haensch, Wilfried
Kim, Seyoung
Neural Network Training With Asymmetric Crosspoint Elements
title Neural Network Training With Asymmetric Crosspoint Elements
title_full Neural Network Training With Asymmetric Crosspoint Elements
title_fullStr Neural Network Training With Asymmetric Crosspoint Elements
title_full_unstemmed Neural Network Training With Asymmetric Crosspoint Elements
title_short Neural Network Training With Asymmetric Crosspoint Elements
title_sort neural network training with asymmetric crosspoint elements
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9124763/
https://www.ncbi.nlm.nih.gov/pubmed/35615470
http://dx.doi.org/10.3389/frai.2022.891624
work_keys_str_mv AT onenmurat neuralnetworktrainingwithasymmetriccrosspointelements
AT gokmentayfun neuralnetworktrainingwithasymmetriccrosspointelements
AT todorovteodork neuralnetworktrainingwithasymmetriccrosspointelements
AT nowickitomasz neuralnetworktrainingwithasymmetriccrosspointelements
AT delalamojesusa neuralnetworktrainingwithasymmetriccrosspointelements
AT rozenjohn neuralnetworktrainingwithasymmetriccrosspointelements
AT haenschwilfried neuralnetworktrainingwithasymmetriccrosspointelements
AT kimseyoung neuralnetworktrainingwithasymmetriccrosspointelements