Cargando…
Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2)
[Image: see text] Machine learning and signal processing on the edge are poised to influence our everyday lives with devices that will learn and infer from data generated by smart sensors and other devices for the Internet of Things. The next leap toward ubiquitous electronics requires increased ene...
Autores principales: | , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
American Chemical Society
2022
|
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8945700/ https://www.ncbi.nlm.nih.gov/pubmed/35167265 http://dx.doi.org/10.1021/acsnano.1c07065 |
_version_ | 1784674016080953344 |
---|---|
author | Migliato Marega, Guilherme Wang, Zhenyu Paliy, Maksym Giusi, Gino Strangio, Sebastiano Castiglione, Francesco Callegari, Christian Tripathi, Mukesh Radenovic, Aleksandra Iannaccone, Giuseppe Kis, Andras |
author_facet | Migliato Marega, Guilherme Wang, Zhenyu Paliy, Maksym Giusi, Gino Strangio, Sebastiano Castiglione, Francesco Callegari, Christian Tripathi, Mukesh Radenovic, Aleksandra Iannaccone, Giuseppe Kis, Andras |
author_sort | Migliato Marega, Guilherme |
collection | PubMed |
description | [Image: see text] Machine learning and signal processing on the edge are poised to influence our everyday lives with devices that will learn and infer from data generated by smart sensors and other devices for the Internet of Things. The next leap toward ubiquitous electronics requires increased energy efficiency of processors for specialized data-driven applications. Here, we show how an in-memory processor fabricated using a two-dimensional materials platform can potentially outperform its silicon counterparts in both standard and nontraditional Von Neumann architectures for artificial neural networks. We have fabricated a flash memory array with a two-dimensional channel using wafer-scale MoS(2). Simulations and experiments show that the device can be scaled down to sub-micrometer channel length without any significant impact on its memory performance and that in simulation a reasonable memory window still exists at sub-50 nm channel lengths. Each device conductance in our circuit can be tuned with a 4-bit precision by closed-loop programming. Using our physical circuit, we demonstrate seven-segment digit display classification with a 91.5% accuracy with training performed ex situ and transferred from a host. Further simulations project that at a system level, the large memory arrays can perform AlexNet classification with an upper limit of 50 000 TOpS/W, potentially outperforming neural network integrated circuits based on double-poly CMOS technology. |
format | Online Article Text |
id | pubmed-8945700 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | American Chemical Society |
record_format | MEDLINE/PubMed |
spelling | pubmed-89457002022-03-28 Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2) Migliato Marega, Guilherme Wang, Zhenyu Paliy, Maksym Giusi, Gino Strangio, Sebastiano Castiglione, Francesco Callegari, Christian Tripathi, Mukesh Radenovic, Aleksandra Iannaccone, Giuseppe Kis, Andras ACS Nano [Image: see text] Machine learning and signal processing on the edge are poised to influence our everyday lives with devices that will learn and infer from data generated by smart sensors and other devices for the Internet of Things. The next leap toward ubiquitous electronics requires increased energy efficiency of processors for specialized data-driven applications. Here, we show how an in-memory processor fabricated using a two-dimensional materials platform can potentially outperform its silicon counterparts in both standard and nontraditional Von Neumann architectures for artificial neural networks. We have fabricated a flash memory array with a two-dimensional channel using wafer-scale MoS(2). Simulations and experiments show that the device can be scaled down to sub-micrometer channel length without any significant impact on its memory performance and that in simulation a reasonable memory window still exists at sub-50 nm channel lengths. Each device conductance in our circuit can be tuned with a 4-bit precision by closed-loop programming. Using our physical circuit, we demonstrate seven-segment digit display classification with a 91.5% accuracy with training performed ex situ and transferred from a host. Further simulations project that at a system level, the large memory arrays can perform AlexNet classification with an upper limit of 50 000 TOpS/W, potentially outperforming neural network integrated circuits based on double-poly CMOS technology. American Chemical Society 2022-02-15 2022-03-22 /pmc/articles/PMC8945700/ /pubmed/35167265 http://dx.doi.org/10.1021/acsnano.1c07065 Text en © 2022 The Authors. Published by American Chemical Society https://creativecommons.org/licenses/by-nc-nd/4.0/Permits non-commercial access and re-use, provided that author attribution and integrity are maintained; but does not permit creation of adaptations or other derivative works (https://creativecommons.org/licenses/by-nc-nd/4.0/). |
spellingShingle | Migliato Marega, Guilherme Wang, Zhenyu Paliy, Maksym Giusi, Gino Strangio, Sebastiano Castiglione, Francesco Callegari, Christian Tripathi, Mukesh Radenovic, Aleksandra Iannaccone, Giuseppe Kis, Andras Low-Power Artificial Neural Network Perceptron Based on Monolayer MoS(2) |
title | Low-Power
Artificial Neural Network Perceptron Based
on Monolayer MoS(2) |
title_full | Low-Power
Artificial Neural Network Perceptron Based
on Monolayer MoS(2) |
title_fullStr | Low-Power
Artificial Neural Network Perceptron Based
on Monolayer MoS(2) |
title_full_unstemmed | Low-Power
Artificial Neural Network Perceptron Based
on Monolayer MoS(2) |
title_short | Low-Power
Artificial Neural Network Perceptron Based
on Monolayer MoS(2) |
title_sort | low-power
artificial neural network perceptron based
on monolayer mos(2) |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8945700/ https://www.ncbi.nlm.nih.gov/pubmed/35167265 http://dx.doi.org/10.1021/acsnano.1c07065 |
work_keys_str_mv | AT migliatomaregaguilherme lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT wangzhenyu lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT paliymaksym lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT giusigino lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT strangiosebastiano lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT castiglionefrancesco lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT callegarichristian lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT tripathimukesh lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT radenovicaleksandra lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT iannacconegiuseppe lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 AT kisandras lowpowerartificialneuralnetworkperceptronbasedonmonolayermos2 |