Computing in Memory
Compute-in-memory (CIM) aims to overcome the von Neumann bottleneck by integrating computation directly within memory, primarily using memristors, to drastically improve energy efficiency and speed in machine learning. Current research focuses on optimizing CIM architectures for various neural network models, including convolutional neural networks (CNNs), transformers, and spiking neural networks (SNNs), often employing techniques like quantization, pruning, and novel training algorithms to address memristor non-idealities. This approach holds significant promise for enabling energy-efficient and high-performance AI at the edge, particularly in resource-constrained applications like mobile devices and embedded systems.
Papers
CIM-MLC: A Multi-level Compilation Stack for Computing-In-Memory Accelerators
Songyun Qu, Shixin Zhao, Bing Li, Yintao He, Xuyi Cai, Lei Zhang, Ying Wang
Enhancing Reliability of Neural Networks at the Edge: Inverted Normalization with Stochastic Affine Transformations
Soyed Tuhin Ahmed, Kamal Danouchi, Guillaume Prenat, Lorena Anghel, Mehdi B. Tahoori