Cargando…

Implementation of Analog Perceptron as an Essential Element of Configurable Neural Networks

Perceptron is an essential element in neural network (NN)-based machine learning, however, the effectiveness of various implementations by circuits is rarely demonstrated from chip testing. This paper presents the measured silicon results for the analog perceptron circuits fabricated in a 0.6 [Formu...

Descripción completa

Detalles Bibliográficos
Autores principales: Geng, Chao, Sun, Qingji, Nakatake, Shigetoshi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7435680/
https://www.ncbi.nlm.nih.gov/pubmed/32751288
http://dx.doi.org/10.3390/s20154222
_version_ 1783572378535591936
author Geng, Chao
Sun, Qingji
Nakatake, Shigetoshi
author_facet Geng, Chao
Sun, Qingji
Nakatake, Shigetoshi
author_sort Geng, Chao
collection PubMed
description Perceptron is an essential element in neural network (NN)-based machine learning, however, the effectiveness of various implementations by circuits is rarely demonstrated from chip testing. This paper presents the measured silicon results for the analog perceptron circuits fabricated in a 0.6 [Formula: see text] m/±2.5 V complementary metal oxide semiconductor (CMOS) process, which are comprised of digital-to-analog converter (DAC)-based multipliers and phase shifters. The results from the measurement convinces us that our implementation attains the correct function and good performance. Furthermore, we propose the multi-layer perceptron (MLP) by utilizing analog perceptron where the structure and neurons as well as weights can be flexibly configured. The example given is to design a 2-3-4 MLP circuit with rectified linear unit (ReLU) activation, which consists of 2 input neurons, 3 hidden neurons, and 4 output neurons. Its experimental case shows that the simulated performance achieves a power dissipation of 200 mW, a range of working frequency from 0 to 1 MHz, and an error ratio within 12.7%. Finally, to demonstrate the feasibility and effectiveness of our analog perceptron for configuring a MLP, seven more analog-based MLPs designed with the same approach are used to analyze the simulation results with respect to various specifications, in which two cases are used to compare to their digital counterparts with the same structures.
format Online
Article
Text
id pubmed-7435680
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-74356802020-08-28 Implementation of Analog Perceptron as an Essential Element of Configurable Neural Networks Geng, Chao Sun, Qingji Nakatake, Shigetoshi Sensors (Basel) Article Perceptron is an essential element in neural network (NN)-based machine learning, however, the effectiveness of various implementations by circuits is rarely demonstrated from chip testing. This paper presents the measured silicon results for the analog perceptron circuits fabricated in a 0.6 [Formula: see text] m/±2.5 V complementary metal oxide semiconductor (CMOS) process, which are comprised of digital-to-analog converter (DAC)-based multipliers and phase shifters. The results from the measurement convinces us that our implementation attains the correct function and good performance. Furthermore, we propose the multi-layer perceptron (MLP) by utilizing analog perceptron where the structure and neurons as well as weights can be flexibly configured. The example given is to design a 2-3-4 MLP circuit with rectified linear unit (ReLU) activation, which consists of 2 input neurons, 3 hidden neurons, and 4 output neurons. Its experimental case shows that the simulated performance achieves a power dissipation of 200 mW, a range of working frequency from 0 to 1 MHz, and an error ratio within 12.7%. Finally, to demonstrate the feasibility and effectiveness of our analog perceptron for configuring a MLP, seven more analog-based MLPs designed with the same approach are used to analyze the simulation results with respect to various specifications, in which two cases are used to compare to their digital counterparts with the same structures. MDPI 2020-07-29 /pmc/articles/PMC7435680/ /pubmed/32751288 http://dx.doi.org/10.3390/s20154222 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Geng, Chao
Sun, Qingji
Nakatake, Shigetoshi
Implementation of Analog Perceptron as an Essential Element of Configurable Neural Networks
title Implementation of Analog Perceptron as an Essential Element of Configurable Neural Networks
title_full Implementation of Analog Perceptron as an Essential Element of Configurable Neural Networks
title_fullStr Implementation of Analog Perceptron as an Essential Element of Configurable Neural Networks
title_full_unstemmed Implementation of Analog Perceptron as an Essential Element of Configurable Neural Networks
title_short Implementation of Analog Perceptron as an Essential Element of Configurable Neural Networks
title_sort implementation of analog perceptron as an essential element of configurable neural networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7435680/
https://www.ncbi.nlm.nih.gov/pubmed/32751288
http://dx.doi.org/10.3390/s20154222
work_keys_str_mv AT gengchao implementationofanalogperceptronasanessentialelementofconfigurableneuralnetworks
AT sunqingji implementationofanalogperceptronasanessentialelementofconfigurableneuralnetworks
AT nakatakeshigetoshi implementationofanalogperceptronasanessentialelementofconfigurableneuralnetworks