Cargando…

Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM †

A repeatable and deterministic non-random weight initialization method in convolutional layers of neural networks examined with the Fast Gradient Sign Method (FSGM). Using the FSGM approach as a technique to measure the initialization effect with controlled distortions in transferred learning, varyi...

Descripción completa

Detalles Bibliográficos
Autores principales: Rudd-Orthner, Richard N. M., Mihaylova, Lyudmila
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8309697/
https://www.ncbi.nlm.nih.gov/pubmed/34300512
http://dx.doi.org/10.3390/s21144772
_version_ 1783728582092128256
author Rudd-Orthner, Richard N. M.
Mihaylova, Lyudmila
author_facet Rudd-Orthner, Richard N. M.
Mihaylova, Lyudmila
author_sort Rudd-Orthner, Richard N. M.
collection PubMed
description A repeatable and deterministic non-random weight initialization method in convolutional layers of neural networks examined with the Fast Gradient Sign Method (FSGM). Using the FSGM approach as a technique to measure the initialization effect with controlled distortions in transferred learning, varying the dataset numerical similarity. The focus is on convolutional layers with induced earlier learning through the use of striped forms for image classification. Which provided a higher performing accuracy in the first epoch, with improvements of between 3–5% in a well known benchmark model, and also ~10% in a color image dataset (MTARSI2), using a dissimilar model architecture. The proposed method is robust to limit optimization approaches like Glorot/Xavier and He initialization. Arguably the approach is within a new category of weight initialization methods, as a number sequence substitution of random numbers, without a tether to the dataset. When examined under the FGSM approach with transferred learning, the proposed method when used with higher distortions (numerically dissimilar datasets), is less compromised against the original cross-validation dataset, at ~31% accuracy instead of ~9%. This is an indication of higher retention of the original fitting in transferred learning.
format Online
Article
Text
id pubmed-8309697
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-83096972021-07-25 Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM † Rudd-Orthner, Richard N. M. Mihaylova, Lyudmila Sensors (Basel) Article A repeatable and deterministic non-random weight initialization method in convolutional layers of neural networks examined with the Fast Gradient Sign Method (FSGM). Using the FSGM approach as a technique to measure the initialization effect with controlled distortions in transferred learning, varying the dataset numerical similarity. The focus is on convolutional layers with induced earlier learning through the use of striped forms for image classification. Which provided a higher performing accuracy in the first epoch, with improvements of between 3–5% in a well known benchmark model, and also ~10% in a color image dataset (MTARSI2), using a dissimilar model architecture. The proposed method is robust to limit optimization approaches like Glorot/Xavier and He initialization. Arguably the approach is within a new category of weight initialization methods, as a number sequence substitution of random numbers, without a tether to the dataset. When examined under the FGSM approach with transferred learning, the proposed method when used with higher distortions (numerically dissimilar datasets), is less compromised against the original cross-validation dataset, at ~31% accuracy instead of ~9%. This is an indication of higher retention of the original fitting in transferred learning. MDPI 2021-07-13 /pmc/articles/PMC8309697/ /pubmed/34300512 http://dx.doi.org/10.3390/s21144772 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Rudd-Orthner, Richard N. M.
Mihaylova, Lyudmila
Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM †
title Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM †
title_full Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM †
title_fullStr Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM †
title_full_unstemmed Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM †
title_short Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM †
title_sort deep convnet: non-random weight initialization for repeatable determinism, examined with fsgm †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8309697/
https://www.ncbi.nlm.nih.gov/pubmed/34300512
http://dx.doi.org/10.3390/s21144772
work_keys_str_mv AT ruddorthnerrichardnm deepconvnetnonrandomweightinitializationforrepeatabledeterminismexaminedwithfsgm
AT mihaylovalyudmila deepconvnetnonrandomweightinitializationforrepeatabledeterminismexaminedwithfsgm