Computing in Memory

Compute-in-memory (CIM) aims to overcome the von Neumann bottleneck by integrating computation directly within memory, primarily using memristors, to drastically improve energy efficiency and speed in machine learning. Current research focuses on optimizing CIM architectures for various neural network models, including convolutional neural networks (CNNs), transformers, and spiking neural networks (SNNs), often employing techniques like quantization, pruning, and novel training algorithms to address memristor non-idealities. This approach holds significant promise for enabling energy-efficient and high-performance AI at the edge, particularly in resource-constrained applications like mobile devices and embedded systems.

Papers