Cargando…

Online Learning for DNN Training: A Stochastic Block Adaptive Gradient Algorithm

Adaptive algorithms are widely used because of their fast convergence rate for training deep neural networks (DNNs). However, the training cost becomes prohibitively expensive due to the computation of the full gradient when training complicated DNN. To reduce the computational cost, we present a st...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Jianghui, Li, Baozhu, Zhou, Yangfan, Zhao, Xuhui, Zhu, Junlong, Zhang, Mingchuan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9184190/
https://www.ncbi.nlm.nih.gov/pubmed/35694581
http://dx.doi.org/10.1155/2022/9337209
Descripción
Sumario:Adaptive algorithms are widely used because of their fast convergence rate for training deep neural networks (DNNs). However, the training cost becomes prohibitively expensive due to the computation of the full gradient when training complicated DNN. To reduce the computational cost, we present a stochastic block adaptive gradient online training algorithm in this study, called SBAG. In this algorithm, stochastic block coordinate descent and the adaptive learning rate are utilized at each iteration. We also prove that the regret bound of [Formula: see text] can be achieved via SBAG, in which T is a time horizon. In addition, we use SBAG to train ResNet-34 and DenseNet-121 on CIFAR-10, respectively. The results demonstrate that SBAG has better training speed and generalized ability than other existing training methods.