Cargando…

Fundamental limits to learning closed-form mathematical models from data

Given a finite and noisy dataset generated with a closed-form mathematical model, when is it possible to learn the true generating model from the data alone? This is the question we investigate here. We show that this model-learning problem displays a transition from a low-noise phase in which the t...

Descripción completa

Detalles Bibliográficos
Autores principales: Fajardo-Fontiveros, Oscar, Reichardt, Ignasi, De Los Ríos, Harry R., Duch, Jordi, Sales-Pardo, Marta, Guimerà, Roger
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9950473/
https://www.ncbi.nlm.nih.gov/pubmed/36823107
http://dx.doi.org/10.1038/s41467-023-36657-z
Descripción
Sumario:Given a finite and noisy dataset generated with a closed-form mathematical model, when is it possible to learn the true generating model from the data alone? This is the question we investigate here. We show that this model-learning problem displays a transition from a low-noise phase in which the true model can be learned, to a phase in which the observation noise is too high for the true model to be learned by any method. Both in the low-noise phase and in the high-noise phase, probabilistic model selection leads to optimal generalization to unseen data. This is in contrast to standard machine learning approaches, including artificial neural networks, which in this particular problem are limited, in the low-noise phase, by their ability to interpolate. In the transition region between the learnable and unlearnable phases, generalization is hard for all approaches including probabilistic model selection.