EconPapers    
Economics at your fingertips  
 

A compute-in-memory chip based on resistive random-access memory

Weier Wan (), Rajkumar Kubendran, Clemens Schaefer, Sukru Burc Eryilmaz, Wenqiang Zhang, Dabin Wu, Stephen Deiss, Priyanka Raina, He Qian, Bin Gao (), Siddharth Joshi (), Huaqiang Wu (), H.-S. Philip Wong () and Gert Cauwenberghs ()
Additional contact information
Weier Wan: Stanford University
Rajkumar Kubendran: University of California San Diego
Clemens Schaefer: University of Notre Dame
Sukru Burc Eryilmaz: Stanford University
Wenqiang Zhang: Tsinghua University
Dabin Wu: Tsinghua University
Stephen Deiss: University of California San Diego
Priyanka Raina: Stanford University
He Qian: Tsinghua University
Bin Gao: Tsinghua University
Siddharth Joshi: University of California San Diego
Huaqiang Wu: Tsinghua University
H.-S. Philip Wong: Stanford University
Gert Cauwenberghs: University of California San Diego

Nature, 2022, vol. 608, issue 7923, 504-512

Abstract: Abstract Realizing increasingly complex artificial intelligence (AI) functionalities directly on edge devices calls for unprecedented energy efficiency of edge hardware. Compute-in-memory (CIM) based on resistive random-access memory (RRAM)1 promises to meet such demand by storing AI model weights in dense, analogue and non-volatile RRAM devices, and by performing AI computation directly within RRAM, thus eliminating power-hungry data movement between separate compute and memory2–5. Although recent studies have demonstrated in-memory matrix-vector multiplication on fully integrated RRAM-CIM hardware6–17, it remains a goal for a RRAM-CIM chip to simultaneously deliver high energy efficiency, versatility to support diverse models and software-comparable accuracy. Although efficiency, versatility and accuracy are all indispensable for broad adoption of the technology, the inter-related trade-offs among them cannot be addressed by isolated improvements on any single abstraction level of the design. Here, by co-optimizing across all hierarchies of the design from algorithms and architecture to circuits and devices, we present NeuRRAM—a RRAM-based CIM chip that simultaneously delivers versatility in reconfiguring CIM cores for diverse model architectures, energy efficiency that is two-times better than previous state-of-the-art RRAM-CIM chips across various computational bit-precisions, and inference accuracy comparable to software models quantized to four-bit weights across various AI tasks, including accuracy of 99.0 percent on MNIST18 and 85.7 percent on CIFAR-1019 image classification, 84.7-percent accuracy on Google speech command recognition20, and a 70-percent reduction in image-reconstruction error on a Bayesian image-recovery task.

Date: 2022
References: Add references at CitEc
Citations: View citations in EconPapers (18)

Downloads: (external link)
https://www.nature.com/articles/s41586-022-04992-8 Abstract (text/html)
Access to the full text of the articles in this series is restricted.

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:nat:nature:v:608:y:2022:i:7923:d:10.1038_s41586-022-04992-8

Ordering information: This journal article can be ordered from
https://www.nature.com/

DOI: 10.1038/s41586-022-04992-8

Access Statistics for this article

Nature is currently edited by Magdalena Skipper

More articles in Nature from Nature
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().

 
Page updated 2025-03-19
Handle: RePEc:nat:nature:v:608:y:2022:i:7923:d:10.1038_s41586-022-04992-8