Cargando…

BioGD: Bio-inspired robust gradient descent

Recent research in machine learning pointed to the core problem of state-of-the-art models which impedes their widespread adoption in different domains. The models’ inability to differentiate between noise and subtle, yet significant variation in data leads to their vulnerability to adversarial pert...

Descripción completa

Detalles Bibliográficos
Autores principales: Kulikovskikh, Ilona, Prokhorov, Sergej, Lipić, Tomislav, Legović, Tarzan, Šmuc, Tomislav
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6611597/
https://www.ncbi.nlm.nih.gov/pubmed/31276469
http://dx.doi.org/10.1371/journal.pone.0219004
_version_ 1783432725179400192
author Kulikovskikh, Ilona
Prokhorov, Sergej
Lipić, Tomislav
Legović, Tarzan
Šmuc, Tomislav
author_facet Kulikovskikh, Ilona
Prokhorov, Sergej
Lipić, Tomislav
Legović, Tarzan
Šmuc, Tomislav
author_sort Kulikovskikh, Ilona
collection PubMed
description Recent research in machine learning pointed to the core problem of state-of-the-art models which impedes their widespread adoption in different domains. The models’ inability to differentiate between noise and subtle, yet significant variation in data leads to their vulnerability to adversarial perturbations that cause wrong predictions with high confidence. The study is aimed at identifying whether the algorithms inspired by biological evolution may achieve better results in cases where brittle robustness properties are highly sensitive to the slight noise. To answer this question, we introduce the new robust gradient descent inspired by the stability and adaptability of biological systems to unknown and changing environments. The proposed optimization technique involves an open-ended adaptation process with regard to two hyperparameters inherited from the generalized Verhulst population growth equation. The hyperparameters increase robustness to adversarial noise by penalizing the degree to which hardly visible changes in gradients impact prediction. The empirical evidence on synthetic and experimental datasets confirmed the viability of the bio-inspired gradient descent and suggested promising directions for future research. The code used for computational experiments is provided in a repository at https://github.com/yukinoi/bio_gradient_descent.
format Online
Article
Text
id pubmed-6611597
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-66115972019-07-12 BioGD: Bio-inspired robust gradient descent Kulikovskikh, Ilona Prokhorov, Sergej Lipić, Tomislav Legović, Tarzan Šmuc, Tomislav PLoS One Research Article Recent research in machine learning pointed to the core problem of state-of-the-art models which impedes their widespread adoption in different domains. The models’ inability to differentiate between noise and subtle, yet significant variation in data leads to their vulnerability to adversarial perturbations that cause wrong predictions with high confidence. The study is aimed at identifying whether the algorithms inspired by biological evolution may achieve better results in cases where brittle robustness properties are highly sensitive to the slight noise. To answer this question, we introduce the new robust gradient descent inspired by the stability and adaptability of biological systems to unknown and changing environments. The proposed optimization technique involves an open-ended adaptation process with regard to two hyperparameters inherited from the generalized Verhulst population growth equation. The hyperparameters increase robustness to adversarial noise by penalizing the degree to which hardly visible changes in gradients impact prediction. The empirical evidence on synthetic and experimental datasets confirmed the viability of the bio-inspired gradient descent and suggested promising directions for future research. The code used for computational experiments is provided in a repository at https://github.com/yukinoi/bio_gradient_descent. Public Library of Science 2019-07-05 /pmc/articles/PMC6611597/ /pubmed/31276469 http://dx.doi.org/10.1371/journal.pone.0219004 Text en © 2019 Kulikovskikh et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Kulikovskikh, Ilona
Prokhorov, Sergej
Lipić, Tomislav
Legović, Tarzan
Šmuc, Tomislav
BioGD: Bio-inspired robust gradient descent
title BioGD: Bio-inspired robust gradient descent
title_full BioGD: Bio-inspired robust gradient descent
title_fullStr BioGD: Bio-inspired robust gradient descent
title_full_unstemmed BioGD: Bio-inspired robust gradient descent
title_short BioGD: Bio-inspired robust gradient descent
title_sort biogd: bio-inspired robust gradient descent
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6611597/
https://www.ncbi.nlm.nih.gov/pubmed/31276469
http://dx.doi.org/10.1371/journal.pone.0219004
work_keys_str_mv AT kulikovskikhilona biogdbioinspiredrobustgradientdescent
AT prokhorovsergej biogdbioinspiredrobustgradientdescent
AT lipictomislav biogdbioinspiredrobustgradientdescent
AT legovictarzan biogdbioinspiredrobustgradientdescent
AT smuctomislav biogdbioinspiredrobustgradientdescent