Cargando…

A Frobenius Norm Regularization Method for Convolutional Kernel Tensors in Neural Networks

The convolutional neural network is a very important model of deep learning. It can help avoid the exploding/vanishing gradient problem and improve the generalizability of a neural network if the singular values of the Jacobian of a layer are bounded around 1 in the training process. We propose a ne...

Descripción completa

Detalles Bibliográficos
Autor principal: Guo, Pei-Chang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9792232/
https://www.ncbi.nlm.nih.gov/pubmed/36579174
http://dx.doi.org/10.1155/2022/3277730
_version_ 1784859590405390336
author Guo, Pei-Chang
author_facet Guo, Pei-Chang
author_sort Guo, Pei-Chang
collection PubMed
description The convolutional neural network is a very important model of deep learning. It can help avoid the exploding/vanishing gradient problem and improve the generalizability of a neural network if the singular values of the Jacobian of a layer are bounded around 1 in the training process. We propose a new Frobenius norm penalty function for a convolutional kernel tensor to let the singular values of the corresponding transformation matrix be bounded around 1. We show how to carry out the gradient-type methods. This provides a potentially useful regularization method for the weights of convolutional layers.
format Online
Article
Text
id pubmed-9792232
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-97922322022-12-27 A Frobenius Norm Regularization Method for Convolutional Kernel Tensors in Neural Networks Guo, Pei-Chang Comput Intell Neurosci Research Article The convolutional neural network is a very important model of deep learning. It can help avoid the exploding/vanishing gradient problem and improve the generalizability of a neural network if the singular values of the Jacobian of a layer are bounded around 1 in the training process. We propose a new Frobenius norm penalty function for a convolutional kernel tensor to let the singular values of the corresponding transformation matrix be bounded around 1. We show how to carry out the gradient-type methods. This provides a potentially useful regularization method for the weights of convolutional layers. Hindawi 2022-08-25 /pmc/articles/PMC9792232/ /pubmed/36579174 http://dx.doi.org/10.1155/2022/3277730 Text en Copyright © 2022 Pei-Chang Guo. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Guo, Pei-Chang
A Frobenius Norm Regularization Method for Convolutional Kernel Tensors in Neural Networks
title A Frobenius Norm Regularization Method for Convolutional Kernel Tensors in Neural Networks
title_full A Frobenius Norm Regularization Method for Convolutional Kernel Tensors in Neural Networks
title_fullStr A Frobenius Norm Regularization Method for Convolutional Kernel Tensors in Neural Networks
title_full_unstemmed A Frobenius Norm Regularization Method for Convolutional Kernel Tensors in Neural Networks
title_short A Frobenius Norm Regularization Method for Convolutional Kernel Tensors in Neural Networks
title_sort frobenius norm regularization method for convolutional kernel tensors in neural networks
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9792232/
https://www.ncbi.nlm.nih.gov/pubmed/36579174
http://dx.doi.org/10.1155/2022/3277730
work_keys_str_mv AT guopeichang afrobeniusnormregularizationmethodforconvolutionalkerneltensorsinneuralnetworks
AT guopeichang frobeniusnormregularizationmethodforconvolutionalkerneltensorsinneuralnetworks