Cargando…
Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm
Recent progress in novel non-volatile memory-based synaptic device technologies and their feasibility for matrix-vector multiplication (MVM) has ignited active research on implementing analog neural network training accelerators with resistive crosspoint arrays. While significant performance boost a...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8770851/ https://www.ncbi.nlm.nih.gov/pubmed/35069098 http://dx.doi.org/10.3389/fnins.2021.767953 |
_version_ | 1784635458920120320 |
---|---|
author | Lee, Chaeun Noh, Kyungmi Ji, Wonjae Gokmen, Tayfun Kim, Seyoung |
author_facet | Lee, Chaeun Noh, Kyungmi Ji, Wonjae Gokmen, Tayfun Kim, Seyoung |
author_sort | Lee, Chaeun |
collection | PubMed |
description | Recent progress in novel non-volatile memory-based synaptic device technologies and their feasibility for matrix-vector multiplication (MVM) has ignited active research on implementing analog neural network training accelerators with resistive crosspoint arrays. While significant performance boost as well as area- and power-efficiency is theoretically predicted, the realization of such analog accelerators is largely limited by non-ideal switching characteristics of crosspoint elements. One of the most performance-limiting non-idealities is the conductance update asymmetry which is known to distort the actual weight change values away from the calculation by error back-propagation and, therefore, significantly deteriorates the neural network training performance. To address this issue by an algorithmic remedy, Tiki-Taka algorithm was proposed and shown to be effective for neural network training with asymmetric devices. However, a systematic analysis to reveal the required asymmetry specification to guarantee the neural network performance has been unexplored. Here, we quantitatively analyze the impact of update asymmetry on the neural network training performance when trained with Tiki-Taka algorithm by exploring the space of asymmetry and hyper-parameters and measuring the classification accuracy. We discover that the update asymmetry level of the auxiliary array affects the way the optimizer takes the importance of previous gradients, whereas that of main array affects the frequency of accepting those gradients. We propose a novel calibration method to find the optimal operating point in terms of device and network parameters. By searching over the hyper-parameter space of Tiki-Taka algorithm using interpolation and Gaussian filtering, we find the optimal hyper-parameters efficiently and reveal the optimal range of asymmetry, namely the asymmetry specification. Finally, we show that the analysis and calibration method be applicable to spiking neural networks. |
format | Online Article Text |
id | pubmed-8770851 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-87708512022-01-21 Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm Lee, Chaeun Noh, Kyungmi Ji, Wonjae Gokmen, Tayfun Kim, Seyoung Front Neurosci Neuroscience Recent progress in novel non-volatile memory-based synaptic device technologies and their feasibility for matrix-vector multiplication (MVM) has ignited active research on implementing analog neural network training accelerators with resistive crosspoint arrays. While significant performance boost as well as area- and power-efficiency is theoretically predicted, the realization of such analog accelerators is largely limited by non-ideal switching characteristics of crosspoint elements. One of the most performance-limiting non-idealities is the conductance update asymmetry which is known to distort the actual weight change values away from the calculation by error back-propagation and, therefore, significantly deteriorates the neural network training performance. To address this issue by an algorithmic remedy, Tiki-Taka algorithm was proposed and shown to be effective for neural network training with asymmetric devices. However, a systematic analysis to reveal the required asymmetry specification to guarantee the neural network performance has been unexplored. Here, we quantitatively analyze the impact of update asymmetry on the neural network training performance when trained with Tiki-Taka algorithm by exploring the space of asymmetry and hyper-parameters and measuring the classification accuracy. We discover that the update asymmetry level of the auxiliary array affects the way the optimizer takes the importance of previous gradients, whereas that of main array affects the frequency of accepting those gradients. We propose a novel calibration method to find the optimal operating point in terms of device and network parameters. By searching over the hyper-parameter space of Tiki-Taka algorithm using interpolation and Gaussian filtering, we find the optimal hyper-parameters efficiently and reveal the optimal range of asymmetry, namely the asymmetry specification. Finally, we show that the analysis and calibration method be applicable to spiking neural networks. Frontiers Media S.A. 2022-01-06 /pmc/articles/PMC8770851/ /pubmed/35069098 http://dx.doi.org/10.3389/fnins.2021.767953 Text en Copyright © 2022 Lee, Noh, Ji, Gokmen and Kim. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Lee, Chaeun Noh, Kyungmi Ji, Wonjae Gokmen, Tayfun Kim, Seyoung Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm |
title | Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm |
title_full | Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm |
title_fullStr | Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm |
title_full_unstemmed | Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm |
title_short | Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm |
title_sort | impact of asymmetric weight update on neural network training with tiki-taka algorithm |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8770851/ https://www.ncbi.nlm.nih.gov/pubmed/35069098 http://dx.doi.org/10.3389/fnins.2021.767953 |
work_keys_str_mv | AT leechaeun impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm AT nohkyungmi impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm AT jiwonjae impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm AT gokmentayfun impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm AT kimseyoung impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm |