Cargando…

Decoupled neural network training with re-computation and weight prediction

To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a n...

Descripción completa

Detalles Bibliográficos
Autores principales: Peng, Jiawei, Xu, Yicheng, Lin, Zhiping, Weng, Zhenyu, Yang, Zishuo, Zhuang, Huiping
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9949630/
https://www.ncbi.nlm.nih.gov/pubmed/36821537
http://dx.doi.org/10.1371/journal.pone.0276427
_version_ 1784892985774702592
author Peng, Jiawei
Xu, Yicheng
Lin, Zhiping
Weng, Zhenyu
Yang, Zishuo
Zhuang, Huiping
author_facet Peng, Jiawei
Xu, Yicheng
Lin, Zhiping
Weng, Zhenyu
Yang, Zishuo
Zhuang, Huiping
author_sort Peng, Jiawei
collection PubMed
description To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a new form of decoupled learning, named decoupled neural network training scheme with re-computation and weight prediction (DTRP) is proposed. In DTRP, a re-computation scheme is adopted to solve the memory explosion problem, and a weight prediction scheme is proposed to deal with the weight delay caused by re-computation. Additionally, a batch compensation scheme is developed, allowing the proposed DTRP to run faster. Theoretical analysis shows that DTRP is guaranteed to converge to crical points under certain conditions. Experiments are conducted by training various convolutional neural networks on several classification datasets, showing comparable or better results than the state-of-the-art methods and BP. These experiments also reveal that adopting the proposed method, the memory explosion problem is effectively solved, and a significant acceleration is achieved.
format Online
Article
Text
id pubmed-9949630
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-99496302023-02-24 Decoupled neural network training with re-computation and weight prediction Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping PLoS One Research Article To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a new form of decoupled learning, named decoupled neural network training scheme with re-computation and weight prediction (DTRP) is proposed. In DTRP, a re-computation scheme is adopted to solve the memory explosion problem, and a weight prediction scheme is proposed to deal with the weight delay caused by re-computation. Additionally, a batch compensation scheme is developed, allowing the proposed DTRP to run faster. Theoretical analysis shows that DTRP is guaranteed to converge to crical points under certain conditions. Experiments are conducted by training various convolutional neural networks on several classification datasets, showing comparable or better results than the state-of-the-art methods and BP. These experiments also reveal that adopting the proposed method, the memory explosion problem is effectively solved, and a significant acceleration is achieved. Public Library of Science 2023-02-23 /pmc/articles/PMC9949630/ /pubmed/36821537 http://dx.doi.org/10.1371/journal.pone.0276427 Text en © 2023 Peng et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Peng, Jiawei
Xu, Yicheng
Lin, Zhiping
Weng, Zhenyu
Yang, Zishuo
Zhuang, Huiping
Decoupled neural network training with re-computation and weight prediction
title Decoupled neural network training with re-computation and weight prediction
title_full Decoupled neural network training with re-computation and weight prediction
title_fullStr Decoupled neural network training with re-computation and weight prediction
title_full_unstemmed Decoupled neural network training with re-computation and weight prediction
title_short Decoupled neural network training with re-computation and weight prediction
title_sort decoupled neural network training with re-computation and weight prediction
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9949630/
https://www.ncbi.nlm.nih.gov/pubmed/36821537
http://dx.doi.org/10.1371/journal.pone.0276427
work_keys_str_mv AT pengjiawei decoupledneuralnetworktrainingwithrecomputationandweightprediction
AT xuyicheng decoupledneuralnetworktrainingwithrecomputationandweightprediction
AT linzhiping decoupledneuralnetworktrainingwithrecomputationandweightprediction
AT wengzhenyu decoupledneuralnetworktrainingwithrecomputationandweightprediction
AT yangzishuo decoupledneuralnetworktrainingwithrecomputationandweightprediction
AT zhuanghuiping decoupledneuralnetworktrainingwithrecomputationandweightprediction