Memory Accelerator
Memory accelerators aim to drastically improve the energy efficiency and speed of deep neural network (DNN) inference by performing computations directly within the memory array, minimizing data movement. Current research focuses on optimizing various DNN architectures, including convolutional neural networks (CNNs), recurrent neural networks (RNNs), and transformers, for in-memory computing using techniques like ternary weight quantization and hardware-aware training. This approach holds significant promise for reducing the power consumption and latency of AI applications, particularly in resource-constrained environments like edge devices and IoT systems.
Papers
October 30, 2024
January 23, 2024
April 22, 2023
March 13, 2023
February 16, 2023
January 19, 2022
January 4, 2022