Cargando…

Fractional-Order Deep Backpropagation Neural Network

In recent years, the research of artificial neural networks based on fractional calculus has attracted much attention. In this paper, we proposed a fractional-order deep backpropagation (BP) neural network model with L(2) regularization. The proposed network was optimized by the fractional gradient...

Descripción completa

Detalles Bibliográficos
Autores principales: Bao, Chunhui, Pu, Yifei, Zhang, Yi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6051328/
https://www.ncbi.nlm.nih.gov/pubmed/30065757
http://dx.doi.org/10.1155/2018/7361628
_version_ 1783340508818440192
author Bao, Chunhui
Pu, Yifei
Zhang, Yi
author_facet Bao, Chunhui
Pu, Yifei
Zhang, Yi
author_sort Bao, Chunhui
collection PubMed
description In recent years, the research of artificial neural networks based on fractional calculus has attracted much attention. In this paper, we proposed a fractional-order deep backpropagation (BP) neural network model with L(2) regularization. The proposed network was optimized by the fractional gradient descent method with Caputo derivative. We also illustrated the necessary conditions for the convergence of the proposed network. The influence of L(2) regularization on the convergence was analyzed with the fractional-order variational method. The experiments have been performed on the MNIST dataset to demonstrate that the proposed network was deterministically convergent and can effectively avoid overfitting.
format Online
Article
Text
id pubmed-6051328
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-60513282018-07-31 Fractional-Order Deep Backpropagation Neural Network Bao, Chunhui Pu, Yifei Zhang, Yi Comput Intell Neurosci Research Article In recent years, the research of artificial neural networks based on fractional calculus has attracted much attention. In this paper, we proposed a fractional-order deep backpropagation (BP) neural network model with L(2) regularization. The proposed network was optimized by the fractional gradient descent method with Caputo derivative. We also illustrated the necessary conditions for the convergence of the proposed network. The influence of L(2) regularization on the convergence was analyzed with the fractional-order variational method. The experiments have been performed on the MNIST dataset to demonstrate that the proposed network was deterministically convergent and can effectively avoid overfitting. Hindawi 2018-07-03 /pmc/articles/PMC6051328/ /pubmed/30065757 http://dx.doi.org/10.1155/2018/7361628 Text en Copyright © 2018 Chunhui Bao et al. https://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Bao, Chunhui
Pu, Yifei
Zhang, Yi
Fractional-Order Deep Backpropagation Neural Network
title Fractional-Order Deep Backpropagation Neural Network
title_full Fractional-Order Deep Backpropagation Neural Network
title_fullStr Fractional-Order Deep Backpropagation Neural Network
title_full_unstemmed Fractional-Order Deep Backpropagation Neural Network
title_short Fractional-Order Deep Backpropagation Neural Network
title_sort fractional-order deep backpropagation neural network
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6051328/
https://www.ncbi.nlm.nih.gov/pubmed/30065757
http://dx.doi.org/10.1155/2018/7361628
work_keys_str_mv AT baochunhui fractionalorderdeepbackpropagationneuralnetwork
AT puyifei fractionalorderdeepbackpropagationneuralnetwork
AT zhangyi fractionalorderdeepbackpropagationneuralnetwork