Cargando…
A compute-in-memory chip based on resistive random-access memory
Realizing increasingly complex artificial intelligence (AI) functionalities directly on edge devices calls for unprecedented energy efficiency of edge hardware. Compute-in-memory (CIM) based on resistive random-access memory (RRAM)(1) promises to meet such demand by storing AI model weights in dense...
Autores principales: | , , , , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9385482/ https://www.ncbi.nlm.nih.gov/pubmed/35978128 http://dx.doi.org/10.1038/s41586-022-04992-8 |
_version_ | 1784769603491069952 |
---|---|
author | Wan, Weier Kubendran, Rajkumar Schaefer, Clemens Eryilmaz, Sukru Burc Zhang, Wenqiang Wu, Dabin Deiss, Stephen Raina, Priyanka Qian, He Gao, Bin Joshi, Siddharth Wu, Huaqiang Wong, H.-S. Philip Cauwenberghs, Gert |
author_facet | Wan, Weier Kubendran, Rajkumar Schaefer, Clemens Eryilmaz, Sukru Burc Zhang, Wenqiang Wu, Dabin Deiss, Stephen Raina, Priyanka Qian, He Gao, Bin Joshi, Siddharth Wu, Huaqiang Wong, H.-S. Philip Cauwenberghs, Gert |
author_sort | Wan, Weier |
collection | PubMed |
description | Realizing increasingly complex artificial intelligence (AI) functionalities directly on edge devices calls for unprecedented energy efficiency of edge hardware. Compute-in-memory (CIM) based on resistive random-access memory (RRAM)(1) promises to meet such demand by storing AI model weights in dense, analogue and non-volatile RRAM devices, and by performing AI computation directly within RRAM, thus eliminating power-hungry data movement between separate compute and memory(2–5). Although recent studies have demonstrated in-memory matrix-vector multiplication on fully integrated RRAM-CIM hardware(6–17), it remains a goal for a RRAM-CIM chip to simultaneously deliver high energy efficiency, versatility to support diverse models and software-comparable accuracy. Although efficiency, versatility and accuracy are all indispensable for broad adoption of the technology, the inter-related trade-offs among them cannot be addressed by isolated improvements on any single abstraction level of the design. Here, by co-optimizing across all hierarchies of the design from algorithms and architecture to circuits and devices, we present NeuRRAM—a RRAM-based CIM chip that simultaneously delivers versatility in reconfiguring CIM cores for diverse model architectures, energy efficiency that is two-times better than previous state-of-the-art RRAM-CIM chips across various computational bit-precisions, and inference accuracy comparable to software models quantized to four-bit weights across various AI tasks, including accuracy of 99.0 percent on MNIST(18) and 85.7 percent on CIFAR-10(19) image classification, 84.7-percent accuracy on Google speech command recognition(20), and a 70-percent reduction in image-reconstruction error on a Bayesian image-recovery task. |
format | Online Article Text |
id | pubmed-9385482 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-93854822022-08-19 A compute-in-memory chip based on resistive random-access memory Wan, Weier Kubendran, Rajkumar Schaefer, Clemens Eryilmaz, Sukru Burc Zhang, Wenqiang Wu, Dabin Deiss, Stephen Raina, Priyanka Qian, He Gao, Bin Joshi, Siddharth Wu, Huaqiang Wong, H.-S. Philip Cauwenberghs, Gert Nature Article Realizing increasingly complex artificial intelligence (AI) functionalities directly on edge devices calls for unprecedented energy efficiency of edge hardware. Compute-in-memory (CIM) based on resistive random-access memory (RRAM)(1) promises to meet such demand by storing AI model weights in dense, analogue and non-volatile RRAM devices, and by performing AI computation directly within RRAM, thus eliminating power-hungry data movement between separate compute and memory(2–5). Although recent studies have demonstrated in-memory matrix-vector multiplication on fully integrated RRAM-CIM hardware(6–17), it remains a goal for a RRAM-CIM chip to simultaneously deliver high energy efficiency, versatility to support diverse models and software-comparable accuracy. Although efficiency, versatility and accuracy are all indispensable for broad adoption of the technology, the inter-related trade-offs among them cannot be addressed by isolated improvements on any single abstraction level of the design. Here, by co-optimizing across all hierarchies of the design from algorithms and architecture to circuits and devices, we present NeuRRAM—a RRAM-based CIM chip that simultaneously delivers versatility in reconfiguring CIM cores for diverse model architectures, energy efficiency that is two-times better than previous state-of-the-art RRAM-CIM chips across various computational bit-precisions, and inference accuracy comparable to software models quantized to four-bit weights across various AI tasks, including accuracy of 99.0 percent on MNIST(18) and 85.7 percent on CIFAR-10(19) image classification, 84.7-percent accuracy on Google speech command recognition(20), and a 70-percent reduction in image-reconstruction error on a Bayesian image-recovery task. Nature Publishing Group UK 2022-08-17 2022 /pmc/articles/PMC9385482/ /pubmed/35978128 http://dx.doi.org/10.1038/s41586-022-04992-8 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Wan, Weier Kubendran, Rajkumar Schaefer, Clemens Eryilmaz, Sukru Burc Zhang, Wenqiang Wu, Dabin Deiss, Stephen Raina, Priyanka Qian, He Gao, Bin Joshi, Siddharth Wu, Huaqiang Wong, H.-S. Philip Cauwenberghs, Gert A compute-in-memory chip based on resistive random-access memory |
title | A compute-in-memory chip based on resistive random-access memory |
title_full | A compute-in-memory chip based on resistive random-access memory |
title_fullStr | A compute-in-memory chip based on resistive random-access memory |
title_full_unstemmed | A compute-in-memory chip based on resistive random-access memory |
title_short | A compute-in-memory chip based on resistive random-access memory |
title_sort | compute-in-memory chip based on resistive random-access memory |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9385482/ https://www.ncbi.nlm.nih.gov/pubmed/35978128 http://dx.doi.org/10.1038/s41586-022-04992-8 |
work_keys_str_mv | AT wanweier acomputeinmemorychipbasedonresistiverandomaccessmemory AT kubendranrajkumar acomputeinmemorychipbasedonresistiverandomaccessmemory AT schaeferclemens acomputeinmemorychipbasedonresistiverandomaccessmemory AT eryilmazsukruburc acomputeinmemorychipbasedonresistiverandomaccessmemory AT zhangwenqiang acomputeinmemorychipbasedonresistiverandomaccessmemory AT wudabin acomputeinmemorychipbasedonresistiverandomaccessmemory AT deissstephen acomputeinmemorychipbasedonresistiverandomaccessmemory AT rainapriyanka acomputeinmemorychipbasedonresistiverandomaccessmemory AT qianhe acomputeinmemorychipbasedonresistiverandomaccessmemory AT gaobin acomputeinmemorychipbasedonresistiverandomaccessmemory AT joshisiddharth acomputeinmemorychipbasedonresistiverandomaccessmemory AT wuhuaqiang acomputeinmemorychipbasedonresistiverandomaccessmemory AT wonghsphilip acomputeinmemorychipbasedonresistiverandomaccessmemory AT cauwenberghsgert acomputeinmemorychipbasedonresistiverandomaccessmemory AT wanweier computeinmemorychipbasedonresistiverandomaccessmemory AT kubendranrajkumar computeinmemorychipbasedonresistiverandomaccessmemory AT schaeferclemens computeinmemorychipbasedonresistiverandomaccessmemory AT eryilmazsukruburc computeinmemorychipbasedonresistiverandomaccessmemory AT zhangwenqiang computeinmemorychipbasedonresistiverandomaccessmemory AT wudabin computeinmemorychipbasedonresistiverandomaccessmemory AT deissstephen computeinmemorychipbasedonresistiverandomaccessmemory AT rainapriyanka computeinmemorychipbasedonresistiverandomaccessmemory AT qianhe computeinmemorychipbasedonresistiverandomaccessmemory AT gaobin computeinmemorychipbasedonresistiverandomaccessmemory AT joshisiddharth computeinmemorychipbasedonresistiverandomaccessmemory AT wuhuaqiang computeinmemorychipbasedonresistiverandomaccessmemory AT wonghsphilip computeinmemorychipbasedonresistiverandomaccessmemory AT cauwenberghsgert computeinmemorychipbasedonresistiverandomaccessmemory |