Cargando…

Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2)

[Image: see text] Machine learning and signal processing on the edge are poised to influence our everyday lives with devices that will learn and infer from data generated by smart sensors and other devices for the Internet of Things. The next leap toward ubiquitous electronics requires increased ene...

Descripción completa

Detalles Bibliográficos
Autores principales: Migliato Marega, Guilherme, Wang, Zhenyu, Paliy, Maksym, Giusi, Gino, Strangio, Sebastiano, Castiglione, Francesco, Callegari, Christian, Tripathi, Mukesh, Radenovic, Aleksandra, Iannaccone, Giuseppe, Kis, Andras
Formato: Online Artículo Texto
Lenguaje:English
Publicado: American Chemical Society 2022
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8945700/
https://www.ncbi.nlm.nih.gov/pubmed/35167265
http://dx.doi.org/10.1021/acsnano.1c07065
_version_ 1784674016080953344
author Migliato Marega, Guilherme
Wang, Zhenyu
Paliy, Maksym
Giusi, Gino
Strangio, Sebastiano
Castiglione, Francesco
Callegari, Christian
Tripathi, Mukesh
Radenovic, Aleksandra
Iannaccone, Giuseppe
Kis, Andras
author_facet Migliato Marega, Guilherme
Wang, Zhenyu
Paliy, Maksym
Giusi, Gino
Strangio, Sebastiano
Castiglione, Francesco
Callegari, Christian
Tripathi, Mukesh
Radenovic, Aleksandra
Iannaccone, Giuseppe
Kis, Andras
author_sort Migliato Marega, Guilherme
collection PubMed
description [Image: see text] Machine learning and signal processing on the edge are poised to influence our everyday lives with devices that will learn and infer from data generated by smart sensors and other devices for the Internet of Things. The next leap toward ubiquitous electronics requires increased energy efficiency of processors for specialized data-driven applications. Here, we show how an in-memory processor fabricated using a two-dimensional materials platform can potentially outperform its silicon counterparts in both standard and nontraditional Von Neumann architectures for artificial neural networks. We have fabricated a flash memory array with a two-dimensional channel using wafer-scale MoS(2). Simulations and experiments show that the device can be scaled down to sub-micrometer channel length without any significant impact on its memory performance and that in simulation a reasonable memory window still exists at sub-50 nm channel lengths. Each device conductance in our circuit can be tuned with a 4-bit precision by closed-loop programming. Using our physical circuit, we demonstrate seven-segment digit display classification with a 91.5% accuracy with training performed ex situ and transferred from a host. Further simulations project that at a system level, the large memory arrays can perform AlexNet classification with an upper limit of 50 000 TOpS/W, potentially outperforming neural network integrated circuits based on double-poly CMOS technology.
format Online
Article
Text
id pubmed-8945700
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher American Chemical Society
record_format MEDLINE/PubMed
spelling pubmed-89457002022-03-28 Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2) Migliato Marega, Guilherme Wang, Zhenyu Paliy, Maksym Giusi, Gino Strangio, Sebastiano Castiglione, Francesco Callegari, Christian Tripathi, Mukesh Radenovic, Aleksandra Iannaccone, Giuseppe Kis, Andras ACS Nano [Image: see text] Machine learning and signal processing on the edge are poised to influence our everyday lives with devices that will learn and infer from data generated by smart sensors and other devices for the Internet of Things. The next leap toward ubiquitous electronics requires increased energy efficiency of processors for specialized data-driven applications. Here, we show how an in-memory processor fabricated using a two-dimensional materials platform can potentially outperform its silicon counterparts in both standard and nontraditional Von Neumann architectures for artificial neural networks. We have fabricated a flash memory array with a two-dimensional channel using wafer-scale MoS(2). Simulations and experiments show that the device can be scaled down to sub-micrometer channel length without any significant impact on its memory performance and that in simulation a reasonable memory window still exists at sub-50 nm channel lengths. Each device conductance in our circuit can be tuned with a 4-bit precision by closed-loop programming. Using our physical circuit, we demonstrate seven-segment digit display classification with a 91.5% accuracy with training performed ex situ and transferred from a host. Further simulations project that at a system level, the large memory arrays can perform AlexNet classification with an upper limit of 50 000 TOpS/W, potentially outperforming neural network integrated circuits based on double-poly CMOS technology. American Chemical Society 2022-02-15 2022-03-22 /pmc/articles/PMC8945700/ /pubmed/35167265 http://dx.doi.org/10.1021/acsnano.1c07065 Text en © 2022 The Authors. Published by American Chemical Society https://creativecommons.org/licenses/by-nc-nd/4.0/Permits non-commercial access and re-use, provided that author attribution and integrity are maintained; but does not permit creation of adaptations or other derivative works (https://creativecommons.org/licenses/by-nc-nd/4.0/).
spellingShingle Migliato Marega, Guilherme
Wang, Zhenyu
Paliy, Maksym
Giusi, Gino
Strangio, Sebastiano
Castiglione, Francesco
Callegari, Christian
Tripathi, Mukesh
Radenovic, Aleksandra
Iannaccone, Giuseppe
Kis, Andras
Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2)
title Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2)
title_full Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2)
title_fullStr Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2)
title_full_unstemmed Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2)
title_short Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2)
title_sort low-power artificial neural network perceptron based on monolayer mos(2)
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8945700/
https://www.ncbi.nlm.nih.gov/pubmed/35167265
http://dx.doi.org/10.1021/acsnano.1c07065
work_keys_str_mv AT migliatomaregaguilherme lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT wangzhenyu lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT paliymaksym lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT giusigino lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT strangiosebastiano lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT castiglionefrancesco lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT callegarichristian lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT tripathimukesh lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT radenovicaleksandra lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT iannacconegiuseppe lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2
AT kisandras lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2