Cargando…

Precision Machine Learning

We explore unique considerations involved in fitting machine learning (ML) models to data with very high precision, as is often required for science applications. We empirically compare various function approximation methods and study how they scale with increasing parameters and data. We find that...

Descripción completa

Detalles Bibliográficos
Autores principales: Michaud, Eric J., Liu, Ziming, Tegmark, Max
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9858077/
https://www.ncbi.nlm.nih.gov/pubmed/36673316
http://dx.doi.org/10.3390/e25010175
_version_ 1784874008965021696
author Michaud, Eric J.
Liu, Ziming
Tegmark, Max
author_facet Michaud, Eric J.
Liu, Ziming
Tegmark, Max
author_sort Michaud, Eric J.
collection PubMed
description We explore unique considerations involved in fitting machine learning (ML) models to data with very high precision, as is often required for science applications. We empirically compare various function approximation methods and study how they scale with increasing parameters and data. We find that neural networks (NNs) can often outperform classical approximation methods on high-dimensional examples, by (we hypothesize) auto-discovering and exploiting modular structures therein. However, neural networks trained with common optimizers are less powerful for low-dimensional cases, which motivates us to study the unique properties of neural network loss landscapes and the corresponding optimization challenges that arise in the high precision regime. To address the optimization issue in low dimensions, we develop training tricks which enable us to train neural networks to extremely low loss, close to the limits allowed by numerical precision.
format Online
Article
Text
id pubmed-9858077
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-98580772023-01-21 Precision Machine Learning Michaud, Eric J. Liu, Ziming Tegmark, Max Entropy (Basel) Article We explore unique considerations involved in fitting machine learning (ML) models to data with very high precision, as is often required for science applications. We empirically compare various function approximation methods and study how they scale with increasing parameters and data. We find that neural networks (NNs) can often outperform classical approximation methods on high-dimensional examples, by (we hypothesize) auto-discovering and exploiting modular structures therein. However, neural networks trained with common optimizers are less powerful for low-dimensional cases, which motivates us to study the unique properties of neural network loss landscapes and the corresponding optimization challenges that arise in the high precision regime. To address the optimization issue in low dimensions, we develop training tricks which enable us to train neural networks to extremely low loss, close to the limits allowed by numerical precision. MDPI 2023-01-15 /pmc/articles/PMC9858077/ /pubmed/36673316 http://dx.doi.org/10.3390/e25010175 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Michaud, Eric J.
Liu, Ziming
Tegmark, Max
Precision Machine Learning
title Precision Machine Learning
title_full Precision Machine Learning
title_fullStr Precision Machine Learning
title_full_unstemmed Precision Machine Learning
title_short Precision Machine Learning
title_sort precision machine learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9858077/
https://www.ncbi.nlm.nih.gov/pubmed/36673316
http://dx.doi.org/10.3390/e25010175
work_keys_str_mv AT michaudericj precisionmachinelearning
AT liuziming precisionmachinelearning
AT tegmarkmax precisionmachinelearning