Cargando…

Optimization of Deep Neural Networks Using SoCs with OpenCL

In the optimization of deep neural networks (DNNs) via evolutionary algorithms (EAs) and the implementation of the training necessary for the creation of the objective function, there is often a trade-off between efficiency and flexibility. Pure software solutions implemented on general-purpose proc...

Descripción completa

Detalles Bibliográficos
Autores principales: Gadea-Gironés, Rafael, Colom-Palero, Ricardo, Herrero-Bosch, Vicente
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982427/
https://www.ncbi.nlm.nih.gov/pubmed/29710875
http://dx.doi.org/10.3390/s18051384
_version_ 1783328239328952320
author Gadea-Gironés, Rafael
Colom-Palero, Ricardo
Herrero-Bosch, Vicente
author_facet Gadea-Gironés, Rafael
Colom-Palero, Ricardo
Herrero-Bosch, Vicente
author_sort Gadea-Gironés, Rafael
collection PubMed
description In the optimization of deep neural networks (DNNs) via evolutionary algorithms (EAs) and the implementation of the training necessary for the creation of the objective function, there is often a trade-off between efficiency and flexibility. Pure software solutions implemented on general-purpose processors tend to be slow because they do not take advantage of the inherent parallelism of these devices, whereas hardware realizations based on heterogeneous platforms (combining central processing units (CPUs), graphics processing units (GPUs) and/or field-programmable gate arrays (FPGAs)) are designed based on different solutions using methodologies supported by different languages and using very different implementation criteria. This paper first presents a study that demonstrates the need for a heterogeneous (CPU-GPU-FPGA) platform to accelerate the optimization of artificial neural networks (ANNs) using genetic algorithms. Second, the paper presents implementations of the calculations related to the individuals evaluated in such an algorithm on different (CPU- and FPGA-based) platforms, but with the same source files written in OpenCL. The implementation of individuals on remote, low-cost FPGA systems on a chip (SoCs) is found to enable the achievement of good efficiency in terms of performance per watt.
format Online
Article
Text
id pubmed-5982427
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-59824272018-06-05 Optimization of Deep Neural Networks Using SoCs with OpenCL Gadea-Gironés, Rafael Colom-Palero, Ricardo Herrero-Bosch, Vicente Sensors (Basel) Article In the optimization of deep neural networks (DNNs) via evolutionary algorithms (EAs) and the implementation of the training necessary for the creation of the objective function, there is often a trade-off between efficiency and flexibility. Pure software solutions implemented on general-purpose processors tend to be slow because they do not take advantage of the inherent parallelism of these devices, whereas hardware realizations based on heterogeneous platforms (combining central processing units (CPUs), graphics processing units (GPUs) and/or field-programmable gate arrays (FPGAs)) are designed based on different solutions using methodologies supported by different languages and using very different implementation criteria. This paper first presents a study that demonstrates the need for a heterogeneous (CPU-GPU-FPGA) platform to accelerate the optimization of artificial neural networks (ANNs) using genetic algorithms. Second, the paper presents implementations of the calculations related to the individuals evaluated in such an algorithm on different (CPU- and FPGA-based) platforms, but with the same source files written in OpenCL. The implementation of individuals on remote, low-cost FPGA systems on a chip (SoCs) is found to enable the achievement of good efficiency in terms of performance per watt. MDPI 2018-04-30 /pmc/articles/PMC5982427/ /pubmed/29710875 http://dx.doi.org/10.3390/s18051384 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Gadea-Gironés, Rafael
Colom-Palero, Ricardo
Herrero-Bosch, Vicente
Optimization of Deep Neural Networks Using SoCs with OpenCL
title Optimization of Deep Neural Networks Using SoCs with OpenCL
title_full Optimization of Deep Neural Networks Using SoCs with OpenCL
title_fullStr Optimization of Deep Neural Networks Using SoCs with OpenCL
title_full_unstemmed Optimization of Deep Neural Networks Using SoCs with OpenCL
title_short Optimization of Deep Neural Networks Using SoCs with OpenCL
title_sort optimization of deep neural networks using socs with opencl
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982427/
https://www.ncbi.nlm.nih.gov/pubmed/29710875
http://dx.doi.org/10.3390/s18051384
work_keys_str_mv AT gadeagironesrafael optimizationofdeepneuralnetworksusingsocswithopencl
AT colompaleroricardo optimizationofdeepneuralnetworksusingsocswithopencl
AT herreroboschvicente optimizationofdeepneuralnetworksusingsocswithopencl