Cargando…
High frequency accuracy and loss data of random neural networks trained on image datasets
Neural Networks (NNs) are increasingly used across scientific domains to extract knowledge from experimental or computational data. An NN is composed of natural or artificial neurons that serve as simple processing units and are interconnected into a model architecture; it acquires knowledge from th...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8749157/ https://www.ncbi.nlm.nih.gov/pubmed/35036484 http://dx.doi.org/10.1016/j.dib.2021.107780 |
_version_ | 1784631167738183680 |
---|---|
author | Rorabaugh, Ariel Keller Caíno-Lores, Silvina Johnston, Travis Taufer, Michela |
author_facet | Rorabaugh, Ariel Keller Caíno-Lores, Silvina Johnston, Travis Taufer, Michela |
author_sort | Rorabaugh, Ariel Keller |
collection | PubMed |
description | Neural Networks (NNs) are increasingly used across scientific domains to extract knowledge from experimental or computational data. An NN is composed of natural or artificial neurons that serve as simple processing units and are interconnected into a model architecture; it acquires knowledge from the environment through a learning process and stores this knowledge in its connections. The learning process is conducted by training. During NN training, the learning process can be tracked by periodically validating the NN and calculating its fitness. The resulting sequence of fitness values (i.e., validation accuracy or validation loss) is called the NN learning curve. The development of tools for NN design requires knowledge of diverse NNs and their complete learning curves. Generally, only final fully-trained fitness values for highly accurate NNs are made available to the community, hampering efforts to develop tools for NN design and leaving unaddressed aspects such as explaining the generation of an NN and reproducing its learning process. Our dataset fills this gap by fully recording the structure, metadata, and complete learning curves for a wide variety of random NNs throughout their training. Our dataset captures the lifespan of 6000 NNs throughout generation, training, and validation stages. It consists of a suite of 6000 tables, each table representing the lifespan of one NN. We generate each NN with randomized parameter values and train it for 40 epochs on one of three diverse image datasets (i.e., CIFAR-100, FashionMNIST, SVHN). We calculate and record each NN’s fitness with high frequency—every half epoch—to capture the evolution of the training and validation process. As a result, for each NN, we record the generated parameter values describing the structure of that NN, the image dataset on which the NN trained, and all loss and accuracy values for the NN every half epoch. We put our dataset to the service of researchers studying NN performance and its evolution throughout training and validation. Statistical methods can be applied to our dataset to analyze the shape of learning curves in diverse NNs, and the relationship between an NN’s structure and its fitness. Additionally, the structural data and metadata that we record enable the reconstruction and reproducibility of the associated NN. |
format | Online Article Text |
id | pubmed-8749157 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-87491572022-01-13 High frequency accuracy and loss data of random neural networks trained on image datasets Rorabaugh, Ariel Keller Caíno-Lores, Silvina Johnston, Travis Taufer, Michela Data Brief Data Article Neural Networks (NNs) are increasingly used across scientific domains to extract knowledge from experimental or computational data. An NN is composed of natural or artificial neurons that serve as simple processing units and are interconnected into a model architecture; it acquires knowledge from the environment through a learning process and stores this knowledge in its connections. The learning process is conducted by training. During NN training, the learning process can be tracked by periodically validating the NN and calculating its fitness. The resulting sequence of fitness values (i.e., validation accuracy or validation loss) is called the NN learning curve. The development of tools for NN design requires knowledge of diverse NNs and their complete learning curves. Generally, only final fully-trained fitness values for highly accurate NNs are made available to the community, hampering efforts to develop tools for NN design and leaving unaddressed aspects such as explaining the generation of an NN and reproducing its learning process. Our dataset fills this gap by fully recording the structure, metadata, and complete learning curves for a wide variety of random NNs throughout their training. Our dataset captures the lifespan of 6000 NNs throughout generation, training, and validation stages. It consists of a suite of 6000 tables, each table representing the lifespan of one NN. We generate each NN with randomized parameter values and train it for 40 epochs on one of three diverse image datasets (i.e., CIFAR-100, FashionMNIST, SVHN). We calculate and record each NN’s fitness with high frequency—every half epoch—to capture the evolution of the training and validation process. As a result, for each NN, we record the generated parameter values describing the structure of that NN, the image dataset on which the NN trained, and all loss and accuracy values for the NN every half epoch. We put our dataset to the service of researchers studying NN performance and its evolution throughout training and validation. Statistical methods can be applied to our dataset to analyze the shape of learning curves in diverse NNs, and the relationship between an NN’s structure and its fitness. Additionally, the structural data and metadata that we record enable the reconstruction and reproducibility of the associated NN. Elsevier 2022-01-05 /pmc/articles/PMC8749157/ /pubmed/35036484 http://dx.doi.org/10.1016/j.dib.2021.107780 Text en © 2022 The Author(s) https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Data Article Rorabaugh, Ariel Keller Caíno-Lores, Silvina Johnston, Travis Taufer, Michela High frequency accuracy and loss data of random neural networks trained on image datasets |
title | High frequency accuracy and loss data of random neural networks trained on image datasets |
title_full | High frequency accuracy and loss data of random neural networks trained on image datasets |
title_fullStr | High frequency accuracy and loss data of random neural networks trained on image datasets |
title_full_unstemmed | High frequency accuracy and loss data of random neural networks trained on image datasets |
title_short | High frequency accuracy and loss data of random neural networks trained on image datasets |
title_sort | high frequency accuracy and loss data of random neural networks trained on image datasets |
topic | Data Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8749157/ https://www.ncbi.nlm.nih.gov/pubmed/35036484 http://dx.doi.org/10.1016/j.dib.2021.107780 |
work_keys_str_mv | AT rorabaugharielkeller highfrequencyaccuracyandlossdataofrandomneuralnetworkstrainedonimagedatasets AT cainoloressilvina highfrequencyaccuracyandlossdataofrandomneuralnetworkstrainedonimagedatasets AT johnstontravis highfrequencyaccuracyandlossdataofrandomneuralnetworkstrainedonimagedatasets AT taufermichela highfrequencyaccuracyandlossdataofrandomneuralnetworkstrainedonimagedatasets |