Cargando…

The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale’s 18th problem

Deep learning (DL) has had unprecedented success and is now entering scientific computing with full force. However, current DL methods typically suffer from instability, even when universal approximation properties guarantee the existence of stable neural networks (NNs). We address this paradox by d...

Descripción completa

Detalles Bibliográficos
Autores principales: Colbrook, Matthew J., Antun, Vegard, Hansen, Anders C.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: National Academy of Sciences 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8944871/
https://www.ncbi.nlm.nih.gov/pubmed/35294283
http://dx.doi.org/10.1073/pnas.2107151119
_version_ 1784673824239779840
author Colbrook, Matthew J.
Antun, Vegard
Hansen, Anders C.
author_facet Colbrook, Matthew J.
Antun, Vegard
Hansen, Anders C.
author_sort Colbrook, Matthew J.
collection PubMed
description Deep learning (DL) has had unprecedented success and is now entering scientific computing with full force. However, current DL methods typically suffer from instability, even when universal approximation properties guarantee the existence of stable neural networks (NNs). We address this paradox by demonstrating basic well-conditioned problems in scientific computing where one can prove the existence of NNs with great approximation qualities; however, there does not exist any algorithm, even randomized, that can train (or compute) such a NN. For any positive integers [Formula: see text] and L, there are cases where simultaneously 1) no randomized training algorithm can compute a NN correct to K digits with probability greater than 1/2; 2) there exists a deterministic training algorithm that computes a NN with K –1 correct digits, but any such (even randomized) algorithm needs arbitrarily many training data; and 3) there exists a deterministic training algorithm that computes a NN with K –2 correct digits using no more than L training samples. These results imply a classification theory describing conditions under which (stable) NNs with a given accuracy can be computed by an algorithm. We begin this theory by establishing sufficient conditions for the existence of algorithms that compute stable NNs in inverse problems. We introduce fast iterative restarted networks (FIRENETs), which we both prove and numerically verify are stable. Moreover, we prove that only [Formula: see text] layers are needed for an ϵ-accurate solution to the inverse problem.
format Online
Article
Text
id pubmed-8944871
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher National Academy of Sciences
record_format MEDLINE/PubMed
spelling pubmed-89448712022-09-16 The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale’s 18th problem Colbrook, Matthew J. Antun, Vegard Hansen, Anders C. Proc Natl Acad Sci U S A Physical Sciences Deep learning (DL) has had unprecedented success and is now entering scientific computing with full force. However, current DL methods typically suffer from instability, even when universal approximation properties guarantee the existence of stable neural networks (NNs). We address this paradox by demonstrating basic well-conditioned problems in scientific computing where one can prove the existence of NNs with great approximation qualities; however, there does not exist any algorithm, even randomized, that can train (or compute) such a NN. For any positive integers [Formula: see text] and L, there are cases where simultaneously 1) no randomized training algorithm can compute a NN correct to K digits with probability greater than 1/2; 2) there exists a deterministic training algorithm that computes a NN with K –1 correct digits, but any such (even randomized) algorithm needs arbitrarily many training data; and 3) there exists a deterministic training algorithm that computes a NN with K –2 correct digits using no more than L training samples. These results imply a classification theory describing conditions under which (stable) NNs with a given accuracy can be computed by an algorithm. We begin this theory by establishing sufficient conditions for the existence of algorithms that compute stable NNs in inverse problems. We introduce fast iterative restarted networks (FIRENETs), which we both prove and numerically verify are stable. Moreover, we prove that only [Formula: see text] layers are needed for an ϵ-accurate solution to the inverse problem. National Academy of Sciences 2022-03-16 2022-03-22 /pmc/articles/PMC8944871/ /pubmed/35294283 http://dx.doi.org/10.1073/pnas.2107151119 Text en Copyright © 2022 the Author(s). Published by PNAS. https://creativecommons.org/licenses/by-nc-nd/4.0/This article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND) (https://creativecommons.org/licenses/by-nc-nd/4.0/) .
spellingShingle Physical Sciences
Colbrook, Matthew J.
Antun, Vegard
Hansen, Anders C.
The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale’s 18th problem
title The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale’s 18th problem
title_full The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale’s 18th problem
title_fullStr The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale’s 18th problem
title_full_unstemmed The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale’s 18th problem
title_short The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale’s 18th problem
title_sort difficulty of computing stable and accurate neural networks: on the barriers of deep learning and smale’s 18th problem
topic Physical Sciences
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8944871/
https://www.ncbi.nlm.nih.gov/pubmed/35294283
http://dx.doi.org/10.1073/pnas.2107151119
work_keys_str_mv AT colbrookmatthewj thedifficultyofcomputingstableandaccurateneuralnetworksonthebarriersofdeeplearningandsmales18thproblem
AT antunvegard thedifficultyofcomputingstableandaccurateneuralnetworksonthebarriersofdeeplearningandsmales18thproblem
AT hansenandersc thedifficultyofcomputingstableandaccurateneuralnetworksonthebarriersofdeeplearningandsmales18thproblem
AT colbrookmatthewj difficultyofcomputingstableandaccurateneuralnetworksonthebarriersofdeeplearningandsmales18thproblem
AT antunvegard difficultyofcomputingstableandaccurateneuralnetworksonthebarriersofdeeplearningandsmales18thproblem
AT hansenandersc difficultyofcomputingstableandaccurateneuralnetworksonthebarriersofdeeplearningandsmales18thproblem