Cargando…
Decoupled neural network training with re-computation and weight prediction
To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a n...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9949630/ https://www.ncbi.nlm.nih.gov/pubmed/36821537 http://dx.doi.org/10.1371/journal.pone.0276427 |
_version_ | 1784892985774702592 |
---|---|
author | Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping |
author_facet | Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping |
author_sort | Peng, Jiawei |
collection | PubMed |
description | To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a new form of decoupled learning, named decoupled neural network training scheme with re-computation and weight prediction (DTRP) is proposed. In DTRP, a re-computation scheme is adopted to solve the memory explosion problem, and a weight prediction scheme is proposed to deal with the weight delay caused by re-computation. Additionally, a batch compensation scheme is developed, allowing the proposed DTRP to run faster. Theoretical analysis shows that DTRP is guaranteed to converge to crical points under certain conditions. Experiments are conducted by training various convolutional neural networks on several classification datasets, showing comparable or better results than the state-of-the-art methods and BP. These experiments also reveal that adopting the proposed method, the memory explosion problem is effectively solved, and a significant acceleration is achieved. |
format | Online Article Text |
id | pubmed-9949630 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-99496302023-02-24 Decoupled neural network training with re-computation and weight prediction Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping PLoS One Research Article To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a new form of decoupled learning, named decoupled neural network training scheme with re-computation and weight prediction (DTRP) is proposed. In DTRP, a re-computation scheme is adopted to solve the memory explosion problem, and a weight prediction scheme is proposed to deal with the weight delay caused by re-computation. Additionally, a batch compensation scheme is developed, allowing the proposed DTRP to run faster. Theoretical analysis shows that DTRP is guaranteed to converge to crical points under certain conditions. Experiments are conducted by training various convolutional neural networks on several classification datasets, showing comparable or better results than the state-of-the-art methods and BP. These experiments also reveal that adopting the proposed method, the memory explosion problem is effectively solved, and a significant acceleration is achieved. Public Library of Science 2023-02-23 /pmc/articles/PMC9949630/ /pubmed/36821537 http://dx.doi.org/10.1371/journal.pone.0276427 Text en © 2023 Peng et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping Decoupled neural network training with re-computation and weight prediction |
title | Decoupled neural network training with re-computation and weight prediction |
title_full | Decoupled neural network training with re-computation and weight prediction |
title_fullStr | Decoupled neural network training with re-computation and weight prediction |
title_full_unstemmed | Decoupled neural network training with re-computation and weight prediction |
title_short | Decoupled neural network training with re-computation and weight prediction |
title_sort | decoupled neural network training with re-computation and weight prediction |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9949630/ https://www.ncbi.nlm.nih.gov/pubmed/36821537 http://dx.doi.org/10.1371/journal.pone.0276427 |
work_keys_str_mv | AT pengjiawei decoupledneuralnetworktrainingwithrecomputationandweightprediction AT xuyicheng decoupledneuralnetworktrainingwithrecomputationandweightprediction AT linzhiping decoupledneuralnetworktrainingwithrecomputationandweightprediction AT wengzhenyu decoupledneuralnetworktrainingwithrecomputationandweightprediction AT yangzishuo decoupledneuralnetworktrainingwithrecomputationandweightprediction AT zhuanghuiping decoupledneuralnetworktrainingwithrecomputationandweightprediction |