Cargando…

SAGRAD: A Program for Neural Network Training with Simulated Annealing and the Conjugate Gradient Method

SAGRAD (Simulated Annealing GRADient), a Fortran 77 program for computing neural networks for classification using batch learning, is discussed. Neural network training in SAGRAD is based on a combination of simulated annealing and Møller’s scaled conjugate gradient algorithm, the latter a variation...

Descripción completa

Detalles Bibliográficos
Autores principales: Bernal, Javier, Torres-Jimenez, Jose
Formato: Online Artículo Texto
Lenguaje:English
Publicado: [Gaithersburg, MD] : U.S. Dept. of Commerce, National Institute of Standards and Technology 2015
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4730672/
https://www.ncbi.nlm.nih.gov/pubmed/26958442
http://dx.doi.org/10.6028/jres.120.009
_version_ 1782412443573551104
author Bernal, Javier
Torres-Jimenez, Jose
author_facet Bernal, Javier
Torres-Jimenez, Jose
author_sort Bernal, Javier
collection PubMed
description SAGRAD (Simulated Annealing GRADient), a Fortran 77 program for computing neural networks for classification using batch learning, is discussed. Neural network training in SAGRAD is based on a combination of simulated annealing and Møller’s scaled conjugate gradient algorithm, the latter a variation of the traditional conjugate gradient method, better suited for the nonquadratic nature of neural networks. Different aspects of the implementation of the training process in SAGRAD are discussed, such as the efficient computation of gradients and multiplication of vectors by Hessian matrices that are required by Møller’s algorithm; the (re)initialization of weights with simulated annealing required to (re)start Møller’s algorithm the first time and each time thereafter that it shows insufficient progress in reaching a possibly local minimum; and the use of simulated annealing when Møller’s algorithm, after possibly making considerable progress, becomes stuck at a local minimum or flat area of weight space. Outlines of the scaled conjugate gradient algorithm, the simulated annealing procedure and the training process used in SAGRAD are presented together with results from running SAGRAD on two examples of training data.
format Online
Article
Text
id pubmed-4730672
institution National Center for Biotechnology Information
language English
publishDate 2015
publisher [Gaithersburg, MD] : U.S. Dept. of Commerce, National Institute of Standards and Technology
record_format MEDLINE/PubMed
spelling pubmed-47306722016-03-08 SAGRAD: A Program for Neural Network Training with Simulated Annealing and the Conjugate Gradient Method Bernal, Javier Torres-Jimenez, Jose J Res Natl Inst Stand Technol Article SAGRAD (Simulated Annealing GRADient), a Fortran 77 program for computing neural networks for classification using batch learning, is discussed. Neural network training in SAGRAD is based on a combination of simulated annealing and Møller’s scaled conjugate gradient algorithm, the latter a variation of the traditional conjugate gradient method, better suited for the nonquadratic nature of neural networks. Different aspects of the implementation of the training process in SAGRAD are discussed, such as the efficient computation of gradients and multiplication of vectors by Hessian matrices that are required by Møller’s algorithm; the (re)initialization of weights with simulated annealing required to (re)start Møller’s algorithm the first time and each time thereafter that it shows insufficient progress in reaching a possibly local minimum; and the use of simulated annealing when Møller’s algorithm, after possibly making considerable progress, becomes stuck at a local minimum or flat area of weight space. Outlines of the scaled conjugate gradient algorithm, the simulated annealing procedure and the training process used in SAGRAD are presented together with results from running SAGRAD on two examples of training data. [Gaithersburg, MD] : U.S. Dept. of Commerce, National Institute of Standards and Technology 2015-06-17 /pmc/articles/PMC4730672/ /pubmed/26958442 http://dx.doi.org/10.6028/jres.120.009 Text en https://creativecommons.org/publicdomain/zero/1.0/ The Journal of Research of the National Institute of Standards and Technology is a publication of the U.S. Government. The papers are in the public domain and are not subject to copyright in the United States. Articles from J Res may contain photographs or illustrations copyrighted by other commercial organizations or individuals that may not be used without obtaining prior approval from the holder of the copyright.
spellingShingle Article
Bernal, Javier
Torres-Jimenez, Jose
SAGRAD: A Program for Neural Network Training with Simulated Annealing and the Conjugate Gradient Method
title SAGRAD: A Program for Neural Network Training with Simulated Annealing and the Conjugate Gradient Method
title_full SAGRAD: A Program for Neural Network Training with Simulated Annealing and the Conjugate Gradient Method
title_fullStr SAGRAD: A Program for Neural Network Training with Simulated Annealing and the Conjugate Gradient Method
title_full_unstemmed SAGRAD: A Program for Neural Network Training with Simulated Annealing and the Conjugate Gradient Method
title_short SAGRAD: A Program for Neural Network Training with Simulated Annealing and the Conjugate Gradient Method
title_sort sagrad: a program for neural network training with simulated annealing and the conjugate gradient method
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4730672/
https://www.ncbi.nlm.nih.gov/pubmed/26958442
http://dx.doi.org/10.6028/jres.120.009
work_keys_str_mv AT bernaljavier sagradaprogramforneuralnetworktrainingwithsimulatedannealingandtheconjugategradientmethod
AT torresjimenezjose sagradaprogramforneuralnetworktrainingwithsimulatedannealingandtheconjugategradientmethod