Cargando…
Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM †
A repeatable and deterministic non-random weight initialization method in convolutional layers of neural networks examined with the Fast Gradient Sign Method (FSGM). Using the FSGM approach as a technique to measure the initialization effect with controlled distortions in transferred learning, varyi...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8309697/ https://www.ncbi.nlm.nih.gov/pubmed/34300512 http://dx.doi.org/10.3390/s21144772 |
_version_ | 1783728582092128256 |
---|---|
author | Rudd-Orthner, Richard N. M. Mihaylova, Lyudmila |
author_facet | Rudd-Orthner, Richard N. M. Mihaylova, Lyudmila |
author_sort | Rudd-Orthner, Richard N. M. |
collection | PubMed |
description | A repeatable and deterministic non-random weight initialization method in convolutional layers of neural networks examined with the Fast Gradient Sign Method (FSGM). Using the FSGM approach as a technique to measure the initialization effect with controlled distortions in transferred learning, varying the dataset numerical similarity. The focus is on convolutional layers with induced earlier learning through the use of striped forms for image classification. Which provided a higher performing accuracy in the first epoch, with improvements of between 3–5% in a well known benchmark model, and also ~10% in a color image dataset (MTARSI2), using a dissimilar model architecture. The proposed method is robust to limit optimization approaches like Glorot/Xavier and He initialization. Arguably the approach is within a new category of weight initialization methods, as a number sequence substitution of random numbers, without a tether to the dataset. When examined under the FGSM approach with transferred learning, the proposed method when used with higher distortions (numerically dissimilar datasets), is less compromised against the original cross-validation dataset, at ~31% accuracy instead of ~9%. This is an indication of higher retention of the original fitting in transferred learning. |
format | Online Article Text |
id | pubmed-8309697 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-83096972021-07-25 Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM † Rudd-Orthner, Richard N. M. Mihaylova, Lyudmila Sensors (Basel) Article A repeatable and deterministic non-random weight initialization method in convolutional layers of neural networks examined with the Fast Gradient Sign Method (FSGM). Using the FSGM approach as a technique to measure the initialization effect with controlled distortions in transferred learning, varying the dataset numerical similarity. The focus is on convolutional layers with induced earlier learning through the use of striped forms for image classification. Which provided a higher performing accuracy in the first epoch, with improvements of between 3–5% in a well known benchmark model, and also ~10% in a color image dataset (MTARSI2), using a dissimilar model architecture. The proposed method is robust to limit optimization approaches like Glorot/Xavier and He initialization. Arguably the approach is within a new category of weight initialization methods, as a number sequence substitution of random numbers, without a tether to the dataset. When examined under the FGSM approach with transferred learning, the proposed method when used with higher distortions (numerically dissimilar datasets), is less compromised against the original cross-validation dataset, at ~31% accuracy instead of ~9%. This is an indication of higher retention of the original fitting in transferred learning. MDPI 2021-07-13 /pmc/articles/PMC8309697/ /pubmed/34300512 http://dx.doi.org/10.3390/s21144772 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Rudd-Orthner, Richard N. M. Mihaylova, Lyudmila Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM † |
title | Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM † |
title_full | Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM † |
title_fullStr | Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM † |
title_full_unstemmed | Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM † |
title_short | Deep ConvNet: Non-Random Weight Initialization for Repeatable Determinism, Examined with FSGM † |
title_sort | deep convnet: non-random weight initialization for repeatable determinism, examined with fsgm † |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8309697/ https://www.ncbi.nlm.nih.gov/pubmed/34300512 http://dx.doi.org/10.3390/s21144772 |
work_keys_str_mv | AT ruddorthnerrichardnm deepconvnetnonrandomweightinitializationforrepeatabledeterminismexaminedwithfsgm AT mihaylovalyudmila deepconvnetnonrandomweightinitializationforrepeatabledeterminismexaminedwithfsgm |