Cargando…
Optimization of Deep Neural Networks Using SoCs with OpenCL
In the optimization of deep neural networks (DNNs) via evolutionary algorithms (EAs) and the implementation of the training necessary for the creation of the objective function, there is often a trade-off between efficiency and flexibility. Pure software solutions implemented on general-purpose proc...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982427/ https://www.ncbi.nlm.nih.gov/pubmed/29710875 http://dx.doi.org/10.3390/s18051384 |
_version_ | 1783328239328952320 |
---|---|
author | Gadea-Gironés, Rafael Colom-Palero, Ricardo Herrero-Bosch, Vicente |
author_facet | Gadea-Gironés, Rafael Colom-Palero, Ricardo Herrero-Bosch, Vicente |
author_sort | Gadea-Gironés, Rafael |
collection | PubMed |
description | In the optimization of deep neural networks (DNNs) via evolutionary algorithms (EAs) and the implementation of the training necessary for the creation of the objective function, there is often a trade-off between efficiency and flexibility. Pure software solutions implemented on general-purpose processors tend to be slow because they do not take advantage of the inherent parallelism of these devices, whereas hardware realizations based on heterogeneous platforms (combining central processing units (CPUs), graphics processing units (GPUs) and/or field-programmable gate arrays (FPGAs)) are designed based on different solutions using methodologies supported by different languages and using very different implementation criteria. This paper first presents a study that demonstrates the need for a heterogeneous (CPU-GPU-FPGA) platform to accelerate the optimization of artificial neural networks (ANNs) using genetic algorithms. Second, the paper presents implementations of the calculations related to the individuals evaluated in such an algorithm on different (CPU- and FPGA-based) platforms, but with the same source files written in OpenCL. The implementation of individuals on remote, low-cost FPGA systems on a chip (SoCs) is found to enable the achievement of good efficiency in terms of performance per watt. |
format | Online Article Text |
id | pubmed-5982427 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-59824272018-06-05 Optimization of Deep Neural Networks Using SoCs with OpenCL Gadea-Gironés, Rafael Colom-Palero, Ricardo Herrero-Bosch, Vicente Sensors (Basel) Article In the optimization of deep neural networks (DNNs) via evolutionary algorithms (EAs) and the implementation of the training necessary for the creation of the objective function, there is often a trade-off between efficiency and flexibility. Pure software solutions implemented on general-purpose processors tend to be slow because they do not take advantage of the inherent parallelism of these devices, whereas hardware realizations based on heterogeneous platforms (combining central processing units (CPUs), graphics processing units (GPUs) and/or field-programmable gate arrays (FPGAs)) are designed based on different solutions using methodologies supported by different languages and using very different implementation criteria. This paper first presents a study that demonstrates the need for a heterogeneous (CPU-GPU-FPGA) platform to accelerate the optimization of artificial neural networks (ANNs) using genetic algorithms. Second, the paper presents implementations of the calculations related to the individuals evaluated in such an algorithm on different (CPU- and FPGA-based) platforms, but with the same source files written in OpenCL. The implementation of individuals on remote, low-cost FPGA systems on a chip (SoCs) is found to enable the achievement of good efficiency in terms of performance per watt. MDPI 2018-04-30 /pmc/articles/PMC5982427/ /pubmed/29710875 http://dx.doi.org/10.3390/s18051384 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Gadea-Gironés, Rafael Colom-Palero, Ricardo Herrero-Bosch, Vicente Optimization of Deep Neural Networks Using SoCs with OpenCL |
title | Optimization of Deep Neural Networks Using SoCs with OpenCL |
title_full | Optimization of Deep Neural Networks Using SoCs with OpenCL |
title_fullStr | Optimization of Deep Neural Networks Using SoCs with OpenCL |
title_full_unstemmed | Optimization of Deep Neural Networks Using SoCs with OpenCL |
title_short | Optimization of Deep Neural Networks Using SoCs with OpenCL |
title_sort | optimization of deep neural networks using socs with opencl |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982427/ https://www.ncbi.nlm.nih.gov/pubmed/29710875 http://dx.doi.org/10.3390/s18051384 |
work_keys_str_mv | AT gadeagironesrafael optimizationofdeepneuralnetworksusingsocswithopencl AT colompaleroricardo optimizationofdeepneuralnetworksusingsocswithopencl AT herreroboschvicente optimizationofdeepneuralnetworksusingsocswithopencl |