Cargando…

Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm

Recent progress in novel non-volatile memory-based synaptic device technologies and their feasibility for matrix-vector multiplication (MVM) has ignited active research on implementing analog neural network training accelerators with resistive crosspoint arrays. While significant performance boost a...

Descripción completa

Detalles Bibliográficos
Autores principales: Lee, Chaeun, Noh, Kyungmi, Ji, Wonjae, Gokmen, Tayfun, Kim, Seyoung
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8770851/
https://www.ncbi.nlm.nih.gov/pubmed/35069098
http://dx.doi.org/10.3389/fnins.2021.767953
_version_ 1784635458920120320
author Lee, Chaeun
Noh, Kyungmi
Ji, Wonjae
Gokmen, Tayfun
Kim, Seyoung
author_facet Lee, Chaeun
Noh, Kyungmi
Ji, Wonjae
Gokmen, Tayfun
Kim, Seyoung
author_sort Lee, Chaeun
collection PubMed
description Recent progress in novel non-volatile memory-based synaptic device technologies and their feasibility for matrix-vector multiplication (MVM) has ignited active research on implementing analog neural network training accelerators with resistive crosspoint arrays. While significant performance boost as well as area- and power-efficiency is theoretically predicted, the realization of such analog accelerators is largely limited by non-ideal switching characteristics of crosspoint elements. One of the most performance-limiting non-idealities is the conductance update asymmetry which is known to distort the actual weight change values away from the calculation by error back-propagation and, therefore, significantly deteriorates the neural network training performance. To address this issue by an algorithmic remedy, Tiki-Taka algorithm was proposed and shown to be effective for neural network training with asymmetric devices. However, a systematic analysis to reveal the required asymmetry specification to guarantee the neural network performance has been unexplored. Here, we quantitatively analyze the impact of update asymmetry on the neural network training performance when trained with Tiki-Taka algorithm by exploring the space of asymmetry and hyper-parameters and measuring the classification accuracy. We discover that the update asymmetry level of the auxiliary array affects the way the optimizer takes the importance of previous gradients, whereas that of main array affects the frequency of accepting those gradients. We propose a novel calibration method to find the optimal operating point in terms of device and network parameters. By searching over the hyper-parameter space of Tiki-Taka algorithm using interpolation and Gaussian filtering, we find the optimal hyper-parameters efficiently and reveal the optimal range of asymmetry, namely the asymmetry specification. Finally, we show that the analysis and calibration method be applicable to spiking neural networks.
format Online
Article
Text
id pubmed-8770851
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-87708512022-01-21 Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm Lee, Chaeun Noh, Kyungmi Ji, Wonjae Gokmen, Tayfun Kim, Seyoung Front Neurosci Neuroscience Recent progress in novel non-volatile memory-based synaptic device technologies and their feasibility for matrix-vector multiplication (MVM) has ignited active research on implementing analog neural network training accelerators with resistive crosspoint arrays. While significant performance boost as well as area- and power-efficiency is theoretically predicted, the realization of such analog accelerators is largely limited by non-ideal switching characteristics of crosspoint elements. One of the most performance-limiting non-idealities is the conductance update asymmetry which is known to distort the actual weight change values away from the calculation by error back-propagation and, therefore, significantly deteriorates the neural network training performance. To address this issue by an algorithmic remedy, Tiki-Taka algorithm was proposed and shown to be effective for neural network training with asymmetric devices. However, a systematic analysis to reveal the required asymmetry specification to guarantee the neural network performance has been unexplored. Here, we quantitatively analyze the impact of update asymmetry on the neural network training performance when trained with Tiki-Taka algorithm by exploring the space of asymmetry and hyper-parameters and measuring the classification accuracy. We discover that the update asymmetry level of the auxiliary array affects the way the optimizer takes the importance of previous gradients, whereas that of main array affects the frequency of accepting those gradients. We propose a novel calibration method to find the optimal operating point in terms of device and network parameters. By searching over the hyper-parameter space of Tiki-Taka algorithm using interpolation and Gaussian filtering, we find the optimal hyper-parameters efficiently and reveal the optimal range of asymmetry, namely the asymmetry specification. Finally, we show that the analysis and calibration method be applicable to spiking neural networks. Frontiers Media S.A. 2022-01-06 /pmc/articles/PMC8770851/ /pubmed/35069098 http://dx.doi.org/10.3389/fnins.2021.767953 Text en Copyright © 2022 Lee, Noh, Ji, Gokmen and Kim. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Lee, Chaeun
Noh, Kyungmi
Ji, Wonjae
Gokmen, Tayfun
Kim, Seyoung
Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm
title Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm
title_full Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm
title_fullStr Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm
title_full_unstemmed Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm
title_short Impact of Asymmetric Weight Update on Neural Network Training With Tiki-Taka Algorithm
title_sort impact of asymmetric weight update on neural network training with tiki-taka algorithm
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8770851/
https://www.ncbi.nlm.nih.gov/pubmed/35069098
http://dx.doi.org/10.3389/fnins.2021.767953
work_keys_str_mv AT leechaeun impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm
AT nohkyungmi impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm
AT jiwonjae impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm
AT gokmentayfun impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm
AT kimseyoung impactofasymmetricweightupdateonneuralnetworktrainingwithtikitakaalgorithm