Cargando…

A compute-in-memory chip based on resistive random-access memory

Realizing increasingly complex artificial intelligence (AI) functionalities directly on edge devices calls for unprecedented energy efficiency of edge hardware. Compute-in-memory (CIM) based on resistive random-access memory (RRAM)(1) promises to meet such demand by storing AI model weights in dense...

Descripción completa

Detalles Bibliográficos
Autores principales: Wan, Weier, Kubendran, Rajkumar, Schaefer, Clemens, Eryilmaz, Sukru Burc, Zhang, Wenqiang, Wu, Dabin, Deiss, Stephen, Raina, Priyanka, Qian, He, Gao, Bin, Joshi, Siddharth, Wu, Huaqiang, Wong, H.-S. Philip, Cauwenberghs, Gert
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9385482/
https://www.ncbi.nlm.nih.gov/pubmed/35978128
http://dx.doi.org/10.1038/s41586-022-04992-8
_version_ 1784769603491069952
author Wan, Weier
Kubendran, Rajkumar
Schaefer, Clemens
Eryilmaz, Sukru Burc
Zhang, Wenqiang
Wu, Dabin
Deiss, Stephen
Raina, Priyanka
Qian, He
Gao, Bin
Joshi, Siddharth
Wu, Huaqiang
Wong, H.-S. Philip
Cauwenberghs, Gert
author_facet Wan, Weier
Kubendran, Rajkumar
Schaefer, Clemens
Eryilmaz, Sukru Burc
Zhang, Wenqiang
Wu, Dabin
Deiss, Stephen
Raina, Priyanka
Qian, He
Gao, Bin
Joshi, Siddharth
Wu, Huaqiang
Wong, H.-S. Philip
Cauwenberghs, Gert
author_sort Wan, Weier
collection PubMed
description Realizing increasingly complex artificial intelligence (AI) functionalities directly on edge devices calls for unprecedented energy efficiency of edge hardware. Compute-in-memory (CIM) based on resistive random-access memory (RRAM)(1) promises to meet such demand by storing AI model weights in dense, analogue and non-volatile RRAM devices, and by performing AI computation directly within RRAM, thus eliminating power-hungry data movement between separate compute and memory(2–5). Although recent studies have demonstrated in-memory matrix-vector multiplication on fully integrated RRAM-CIM hardware(6–17), it remains a goal for a RRAM-CIM chip to simultaneously deliver high energy efficiency, versatility to support diverse models and software-comparable accuracy. Although efficiency, versatility and accuracy are all indispensable for broad adoption of the technology, the inter-related trade-offs among them cannot be addressed by isolated improvements on any single abstraction level of the design. Here, by co-optimizing across all hierarchies of the design from algorithms and architecture to circuits and devices, we present NeuRRAM—a RRAM-based CIM chip that simultaneously delivers versatility in reconfiguring CIM cores for diverse model architectures, energy efficiency that is two-times better than previous state-of-the-art RRAM-CIM chips across various computational bit-precisions, and inference accuracy comparable to software models quantized to four-bit weights across various AI tasks, including accuracy of 99.0 percent on MNIST(18) and 85.7 percent on CIFAR-10(19) image classification, 84.7-percent accuracy on Google speech command recognition(20), and a 70-percent reduction in image-reconstruction error on a Bayesian image-recovery task.
format Online
Article
Text
id pubmed-9385482
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-93854822022-08-19 A compute-in-memory chip based on resistive random-access memory Wan, Weier Kubendran, Rajkumar Schaefer, Clemens Eryilmaz, Sukru Burc Zhang, Wenqiang Wu, Dabin Deiss, Stephen Raina, Priyanka Qian, He Gao, Bin Joshi, Siddharth Wu, Huaqiang Wong, H.-S. Philip Cauwenberghs, Gert Nature Article Realizing increasingly complex artificial intelligence (AI) functionalities directly on edge devices calls for unprecedented energy efficiency of edge hardware. Compute-in-memory (CIM) based on resistive random-access memory (RRAM)(1) promises to meet such demand by storing AI model weights in dense, analogue and non-volatile RRAM devices, and by performing AI computation directly within RRAM, thus eliminating power-hungry data movement between separate compute and memory(2–5). Although recent studies have demonstrated in-memory matrix-vector multiplication on fully integrated RRAM-CIM hardware(6–17), it remains a goal for a RRAM-CIM chip to simultaneously deliver high energy efficiency, versatility to support diverse models and software-comparable accuracy. Although efficiency, versatility and accuracy are all indispensable for broad adoption of the technology, the inter-related trade-offs among them cannot be addressed by isolated improvements on any single abstraction level of the design. Here, by co-optimizing across all hierarchies of the design from algorithms and architecture to circuits and devices, we present NeuRRAM—a RRAM-based CIM chip that simultaneously delivers versatility in reconfiguring CIM cores for diverse model architectures, energy efficiency that is two-times better than previous state-of-the-art RRAM-CIM chips across various computational bit-precisions, and inference accuracy comparable to software models quantized to four-bit weights across various AI tasks, including accuracy of 99.0 percent on MNIST(18) and 85.7 percent on CIFAR-10(19) image classification, 84.7-percent accuracy on Google speech command recognition(20), and a 70-percent reduction in image-reconstruction error on a Bayesian image-recovery task. Nature Publishing Group UK 2022-08-17 2022 /pmc/articles/PMC9385482/ /pubmed/35978128 http://dx.doi.org/10.1038/s41586-022-04992-8 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Wan, Weier
Kubendran, Rajkumar
Schaefer, Clemens
Eryilmaz, Sukru Burc
Zhang, Wenqiang
Wu, Dabin
Deiss, Stephen
Raina, Priyanka
Qian, He
Gao, Bin
Joshi, Siddharth
Wu, Huaqiang
Wong, H.-S. Philip
Cauwenberghs, Gert
A compute-in-memory chip based on resistive random-access memory
title A compute-in-memory chip based on resistive random-access memory
title_full A compute-in-memory chip based on resistive random-access memory
title_fullStr A compute-in-memory chip based on resistive random-access memory
title_full_unstemmed A compute-in-memory chip based on resistive random-access memory
title_short A compute-in-memory chip based on resistive random-access memory
title_sort compute-in-memory chip based on resistive random-access memory
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9385482/
https://www.ncbi.nlm.nih.gov/pubmed/35978128
http://dx.doi.org/10.1038/s41586-022-04992-8
work_keys_str_mv AT wanweier acomputeinmemorychipbasedonresistiverandomaccessmemory
AT kubendranrajkumar acomputeinmemorychipbasedonresistiverandomaccessmemory
AT schaeferclemens acomputeinmemorychipbasedonresistiverandomaccessmemory
AT eryilmazsukruburc acomputeinmemorychipbasedonresistiverandomaccessmemory
AT zhangwenqiang acomputeinmemorychipbasedonresistiverandomaccessmemory
AT wudabin acomputeinmemorychipbasedonresistiverandomaccessmemory
AT deissstephen acomputeinmemorychipbasedonresistiverandomaccessmemory
AT rainapriyanka acomputeinmemorychipbasedonresistiverandomaccessmemory
AT qianhe acomputeinmemorychipbasedonresistiverandomaccessmemory
AT gaobin acomputeinmemorychipbasedonresistiverandomaccessmemory
AT joshisiddharth acomputeinmemorychipbasedonresistiverandomaccessmemory
AT wuhuaqiang acomputeinmemorychipbasedonresistiverandomaccessmemory
AT wonghsphilip acomputeinmemorychipbasedonresistiverandomaccessmemory
AT cauwenberghsgert acomputeinmemorychipbasedonresistiverandomaccessmemory
AT wanweier computeinmemorychipbasedonresistiverandomaccessmemory
AT kubendranrajkumar computeinmemorychipbasedonresistiverandomaccessmemory
AT schaeferclemens computeinmemorychipbasedonresistiverandomaccessmemory
AT eryilmazsukruburc computeinmemorychipbasedonresistiverandomaccessmemory
AT zhangwenqiang computeinmemorychipbasedonresistiverandomaccessmemory
AT wudabin computeinmemorychipbasedonresistiverandomaccessmemory
AT deissstephen computeinmemorychipbasedonresistiverandomaccessmemory
AT rainapriyanka computeinmemorychipbasedonresistiverandomaccessmemory
AT qianhe computeinmemorychipbasedonresistiverandomaccessmemory
AT gaobin computeinmemorychipbasedonresistiverandomaccessmemory
AT joshisiddharth computeinmemorychipbasedonresistiverandomaccessmemory
AT wuhuaqiang computeinmemorychipbasedonresistiverandomaccessmemory
AT wonghsphilip computeinmemorychipbasedonresistiverandomaccessmemory
AT cauwenberghsgert computeinmemorychipbasedonresistiverandomaccessmemory