Cargando…

Bayesian Multi-objective Hyperparameter Optimization for Accurate, Fast, and Efficient Neural Network Accelerator Design

In resource-constrained environments, such as low-power edge devices and smart sensors, deploying a fast, compact, and accurate intelligent system with minimum energy is indispensable. Embedding intelligence can be achieved using neural networks on neuromorphic hardware. Designing such networks woul...

Descripción completa

Detalles Bibliográficos
Autores principales: Parsa, Maryam, Mitchell, John P., Schuman, Catherine D., Patton, Robert M., Potok, Thomas E., Roy, Kaushik
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7396641/
https://www.ncbi.nlm.nih.gov/pubmed/32848531
http://dx.doi.org/10.3389/fnins.2020.00667
_version_ 1783565629526114304
author Parsa, Maryam
Mitchell, John P.
Schuman, Catherine D.
Patton, Robert M.
Potok, Thomas E.
Roy, Kaushik
author_facet Parsa, Maryam
Mitchell, John P.
Schuman, Catherine D.
Patton, Robert M.
Potok, Thomas E.
Roy, Kaushik
author_sort Parsa, Maryam
collection PubMed
description In resource-constrained environments, such as low-power edge devices and smart sensors, deploying a fast, compact, and accurate intelligent system with minimum energy is indispensable. Embedding intelligence can be achieved using neural networks on neuromorphic hardware. Designing such networks would require determining several inherent hyperparameters. A key challenge is to find the optimum set of hyperparameters that might belong to the input/output encoding modules, the neural network itself, the application, or the underlying hardware. In this work, we present a hierarchical pseudo agent-based multi-objective Bayesian hyperparameter optimization framework (both software and hardware) that not only maximizes the performance of the network, but also minimizes the energy and area requirements of the corresponding neuromorphic hardware. We validate performance of our approach (in terms of accuracy and computation speed) on several control and classification applications on digital and mixed-signal (memristor-based) neural accelerators. We show that the optimum set of hyperparameters might drastically improve the performance of one application (i.e., 52–71% for Pole-Balance), while having minimum effect on another (i.e., 50–53% for RoboNav). In addition, we demonstrate resiliency of different input/output encoding, training neural network, or the underlying accelerator modules in a neuromorphic system to the changes of the hyperparameters.
format Online
Article
Text
id pubmed-7396641
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-73966412020-08-25 Bayesian Multi-objective Hyperparameter Optimization for Accurate, Fast, and Efficient Neural Network Accelerator Design Parsa, Maryam Mitchell, John P. Schuman, Catherine D. Patton, Robert M. Potok, Thomas E. Roy, Kaushik Front Neurosci Neuroscience In resource-constrained environments, such as low-power edge devices and smart sensors, deploying a fast, compact, and accurate intelligent system with minimum energy is indispensable. Embedding intelligence can be achieved using neural networks on neuromorphic hardware. Designing such networks would require determining several inherent hyperparameters. A key challenge is to find the optimum set of hyperparameters that might belong to the input/output encoding modules, the neural network itself, the application, or the underlying hardware. In this work, we present a hierarchical pseudo agent-based multi-objective Bayesian hyperparameter optimization framework (both software and hardware) that not only maximizes the performance of the network, but also minimizes the energy and area requirements of the corresponding neuromorphic hardware. We validate performance of our approach (in terms of accuracy and computation speed) on several control and classification applications on digital and mixed-signal (memristor-based) neural accelerators. We show that the optimum set of hyperparameters might drastically improve the performance of one application (i.e., 52–71% for Pole-Balance), while having minimum effect on another (i.e., 50–53% for RoboNav). In addition, we demonstrate resiliency of different input/output encoding, training neural network, or the underlying accelerator modules in a neuromorphic system to the changes of the hyperparameters. Frontiers Media S.A. 2020-07-21 /pmc/articles/PMC7396641/ /pubmed/32848531 http://dx.doi.org/10.3389/fnins.2020.00667 Text en Copyright © 2020 Parsa, Mitchell, Schuman, Patton, Potok and Roy. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Parsa, Maryam
Mitchell, John P.
Schuman, Catherine D.
Patton, Robert M.
Potok, Thomas E.
Roy, Kaushik
Bayesian Multi-objective Hyperparameter Optimization for Accurate, Fast, and Efficient Neural Network Accelerator Design
title Bayesian Multi-objective Hyperparameter Optimization for Accurate, Fast, and Efficient Neural Network Accelerator Design
title_full Bayesian Multi-objective Hyperparameter Optimization for Accurate, Fast, and Efficient Neural Network Accelerator Design
title_fullStr Bayesian Multi-objective Hyperparameter Optimization for Accurate, Fast, and Efficient Neural Network Accelerator Design
title_full_unstemmed Bayesian Multi-objective Hyperparameter Optimization for Accurate, Fast, and Efficient Neural Network Accelerator Design
title_short Bayesian Multi-objective Hyperparameter Optimization for Accurate, Fast, and Efficient Neural Network Accelerator Design
title_sort bayesian multi-objective hyperparameter optimization for accurate, fast, and efficient neural network accelerator design
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7396641/
https://www.ncbi.nlm.nih.gov/pubmed/32848531
http://dx.doi.org/10.3389/fnins.2020.00667
work_keys_str_mv AT parsamaryam bayesianmultiobjectivehyperparameteroptimizationforaccuratefastandefficientneuralnetworkacceleratordesign
AT mitchelljohnp bayesianmultiobjectivehyperparameteroptimizationforaccuratefastandefficientneuralnetworkacceleratordesign
AT schumancatherined bayesianmultiobjectivehyperparameteroptimizationforaccuratefastandefficientneuralnetworkacceleratordesign
AT pattonrobertm bayesianmultiobjectivehyperparameteroptimizationforaccuratefastandefficientneuralnetworkacceleratordesign
AT potokthomase bayesianmultiobjectivehyperparameteroptimizationforaccuratefastandefficientneuralnetworkacceleratordesign
AT roykaushik bayesianmultiobjectivehyperparameteroptimizationforaccuratefastandefficientneuralnetworkacceleratordesign