Cargando…

Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn

Neuroscience models commonly have a high number of degrees of freedom and only specific regions within the parameter space are able to produce dynamics of interest. This makes the development of tools and strategies to efficiently find these regions of high importance to advance brain research. Expl...

Descripción completa

Detalles Bibliográficos
Autores principales: Yegenoglu, Alper, Subramoney, Anand, Hater, Thorsten, Jimenez-Romero, Cristian, Klijn, Wouter, Pérez Martín, Aarón, van der Vlag, Michiel, Herty, Michael, Morrison, Abigail, Diaz-Pier, Sandra
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9199579/
https://www.ncbi.nlm.nih.gov/pubmed/35720775
http://dx.doi.org/10.3389/fncom.2022.885207
_version_ 1784727870720966656
author Yegenoglu, Alper
Subramoney, Anand
Hater, Thorsten
Jimenez-Romero, Cristian
Klijn, Wouter
Pérez Martín, Aarón
van der Vlag, Michiel
Herty, Michael
Morrison, Abigail
Diaz-Pier, Sandra
author_facet Yegenoglu, Alper
Subramoney, Anand
Hater, Thorsten
Jimenez-Romero, Cristian
Klijn, Wouter
Pérez Martín, Aarón
van der Vlag, Michiel
Herty, Michael
Morrison, Abigail
Diaz-Pier, Sandra
author_sort Yegenoglu, Alper
collection PubMed
description Neuroscience models commonly have a high number of degrees of freedom and only specific regions within the parameter space are able to produce dynamics of interest. This makes the development of tools and strategies to efficiently find these regions of high importance to advance brain research. Exploring the high dimensional parameter space using numerical simulations has been a frequently used technique in the last years in many areas of computational neuroscience. Today, high performance computing (HPC) can provide a powerful infrastructure to speed up explorations and increase our general understanding of the behavior of the model in reasonable times. Learning to learn (L2L) is a well-known concept in machine learning (ML) and a specific method for acquiring constraints to improve learning performance. This concept can be decomposed into a two loop optimization process where the target of optimization can consist of any program such as an artificial neural network, a spiking network, a single cell model, or a whole brain simulation. In this work, we present L2L as an easy to use and flexible framework to perform parameter and hyper-parameter space exploration of neuroscience models on HPC infrastructure. Learning to learn is an implementation of the L2L concept written in Python. This open-source software allows several instances of an optimization target to be executed with different parameters in an embarrassingly parallel fashion on HPC. L2L provides a set of built-in optimizer algorithms, which make adaptive and efficient exploration of parameter spaces possible. Different from other optimization toolboxes, L2L provides maximum flexibility for the way the optimization target can be executed. In this paper, we show a variety of examples of neuroscience models being optimized within the L2L framework to execute different types of tasks. The tasks used to illustrate the concept go from reproducing empirical data to learning how to solve a problem in a dynamic environment. We particularly focus on simulations with models ranging from the single cell to the whole brain and using a variety of simulation engines like NEST, Arbor, TVB, OpenAIGym, and NetLogo.
format Online
Article
Text
id pubmed-9199579
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-91995792022-06-16 Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn Yegenoglu, Alper Subramoney, Anand Hater, Thorsten Jimenez-Romero, Cristian Klijn, Wouter Pérez Martín, Aarón van der Vlag, Michiel Herty, Michael Morrison, Abigail Diaz-Pier, Sandra Front Comput Neurosci Neuroscience Neuroscience models commonly have a high number of degrees of freedom and only specific regions within the parameter space are able to produce dynamics of interest. This makes the development of tools and strategies to efficiently find these regions of high importance to advance brain research. Exploring the high dimensional parameter space using numerical simulations has been a frequently used technique in the last years in many areas of computational neuroscience. Today, high performance computing (HPC) can provide a powerful infrastructure to speed up explorations and increase our general understanding of the behavior of the model in reasonable times. Learning to learn (L2L) is a well-known concept in machine learning (ML) and a specific method for acquiring constraints to improve learning performance. This concept can be decomposed into a two loop optimization process where the target of optimization can consist of any program such as an artificial neural network, a spiking network, a single cell model, or a whole brain simulation. In this work, we present L2L as an easy to use and flexible framework to perform parameter and hyper-parameter space exploration of neuroscience models on HPC infrastructure. Learning to learn is an implementation of the L2L concept written in Python. This open-source software allows several instances of an optimization target to be executed with different parameters in an embarrassingly parallel fashion on HPC. L2L provides a set of built-in optimizer algorithms, which make adaptive and efficient exploration of parameter spaces possible. Different from other optimization toolboxes, L2L provides maximum flexibility for the way the optimization target can be executed. In this paper, we show a variety of examples of neuroscience models being optimized within the L2L framework to execute different types of tasks. The tasks used to illustrate the concept go from reproducing empirical data to learning how to solve a problem in a dynamic environment. We particularly focus on simulations with models ranging from the single cell to the whole brain and using a variety of simulation engines like NEST, Arbor, TVB, OpenAIGym, and NetLogo. Frontiers Media S.A. 2022-05-27 /pmc/articles/PMC9199579/ /pubmed/35720775 http://dx.doi.org/10.3389/fncom.2022.885207 Text en Copyright © 2022 Yegenoglu, Subramoney, Hater, Jimenez-Romero, Klijn, Pérez Martín, van der Vlag, Herty, Morrison and Diaz-Pier. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Yegenoglu, Alper
Subramoney, Anand
Hater, Thorsten
Jimenez-Romero, Cristian
Klijn, Wouter
Pérez Martín, Aarón
van der Vlag, Michiel
Herty, Michael
Morrison, Abigail
Diaz-Pier, Sandra
Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn
title Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn
title_full Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn
title_fullStr Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn
title_full_unstemmed Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn
title_short Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn
title_sort exploring parameter and hyper-parameter spaces of neuroscience models on high performance computers with learning to learn
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9199579/
https://www.ncbi.nlm.nih.gov/pubmed/35720775
http://dx.doi.org/10.3389/fncom.2022.885207
work_keys_str_mv AT yegenoglualper exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn
AT subramoneyanand exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn
AT haterthorsten exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn
AT jimenezromerocristian exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn
AT klijnwouter exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn
AT perezmartinaaron exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn
AT vandervlagmichiel exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn
AT hertymichael exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn
AT morrisonabigail exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn
AT diazpiersandra exploringparameterandhyperparameterspacesofneurosciencemodelsonhighperformancecomputerswithlearningtolearn