Cargando…

Symbolic expression generation via variational auto-encoder

There are many problems in physics, biology, and other natural sciences in which symbolic regression can provide valuable insights and discover new laws of nature. Widespread deep neural networks do not provide interpretable solutions. Meanwhile, symbolic expressions give us a clear relation between...

Descripción completa

Detalles Bibliográficos
Autores principales: Popov, Sergei, Lazarev, Mikhail, Belavin, Vladislav, Derkach, Denis, Ustyuzhanin, Andrey
Formato: Online Artículo Texto
Lenguaje:English
Publicado: PeerJ Inc. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10280571/
https://www.ncbi.nlm.nih.gov/pubmed/37346583
http://dx.doi.org/10.7717/peerj-cs.1241
_version_ 1785060825175687168
author Popov, Sergei
Lazarev, Mikhail
Belavin, Vladislav
Derkach, Denis
Ustyuzhanin, Andrey
author_facet Popov, Sergei
Lazarev, Mikhail
Belavin, Vladislav
Derkach, Denis
Ustyuzhanin, Andrey
author_sort Popov, Sergei
collection PubMed
description There are many problems in physics, biology, and other natural sciences in which symbolic regression can provide valuable insights and discover new laws of nature. Widespread deep neural networks do not provide interpretable solutions. Meanwhile, symbolic expressions give us a clear relation between observations and the target variable. However, at the moment, there is no dominant solution for the symbolic regression task, and we aim to reduce this gap with our algorithm. In this work, we propose a novel deep learning framework for symbolic expression generation via variational autoencoder (VAE). We suggest using a VAE to generate mathematical expressions, and our training strategy forces generated formulas to fit a given dataset. Our framework allows encoding apriori knowledge of the formulas into fast-check predicates that speed up the optimization process. We compare our method to modern symbolic regression benchmarks and show that our method outperforms the competitors under noisy conditions. The recovery rate of SEGVAE is 65% on the Ngyuen dataset with a noise level of 10%, which is better than the previously reported SOTA by 20%. We demonstrate that this value depends on the dataset and can be even higher.
format Online
Article
Text
id pubmed-10280571
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher PeerJ Inc.
record_format MEDLINE/PubMed
spelling pubmed-102805712023-06-21 Symbolic expression generation via variational auto-encoder Popov, Sergei Lazarev, Mikhail Belavin, Vladislav Derkach, Denis Ustyuzhanin, Andrey PeerJ Comput Sci Artificial Intelligence There are many problems in physics, biology, and other natural sciences in which symbolic regression can provide valuable insights and discover new laws of nature. Widespread deep neural networks do not provide interpretable solutions. Meanwhile, symbolic expressions give us a clear relation between observations and the target variable. However, at the moment, there is no dominant solution for the symbolic regression task, and we aim to reduce this gap with our algorithm. In this work, we propose a novel deep learning framework for symbolic expression generation via variational autoencoder (VAE). We suggest using a VAE to generate mathematical expressions, and our training strategy forces generated formulas to fit a given dataset. Our framework allows encoding apriori knowledge of the formulas into fast-check predicates that speed up the optimization process. We compare our method to modern symbolic regression benchmarks and show that our method outperforms the competitors under noisy conditions. The recovery rate of SEGVAE is 65% on the Ngyuen dataset with a noise level of 10%, which is better than the previously reported SOTA by 20%. We demonstrate that this value depends on the dataset and can be even higher. PeerJ Inc. 2023-03-07 /pmc/articles/PMC10280571/ /pubmed/37346583 http://dx.doi.org/10.7717/peerj-cs.1241 Text en © 2023 Popov et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
spellingShingle Artificial Intelligence
Popov, Sergei
Lazarev, Mikhail
Belavin, Vladislav
Derkach, Denis
Ustyuzhanin, Andrey
Symbolic expression generation via variational auto-encoder
title Symbolic expression generation via variational auto-encoder
title_full Symbolic expression generation via variational auto-encoder
title_fullStr Symbolic expression generation via variational auto-encoder
title_full_unstemmed Symbolic expression generation via variational auto-encoder
title_short Symbolic expression generation via variational auto-encoder
title_sort symbolic expression generation via variational auto-encoder
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10280571/
https://www.ncbi.nlm.nih.gov/pubmed/37346583
http://dx.doi.org/10.7717/peerj-cs.1241
work_keys_str_mv AT popovsergei symbolicexpressiongenerationviavariationalautoencoder
AT lazarevmikhail symbolicexpressiongenerationviavariationalautoencoder
AT belavinvladislav symbolicexpressiongenerationviavariationalautoencoder
AT derkachdenis symbolicexpressiongenerationviavariationalautoencoder
AT ustyuzhaninandrey symbolicexpressiongenerationviavariationalautoencoder