Accelerator Architecture
Accelerator architecture research focuses on designing efficient hardware for executing computationally intensive machine learning models, primarily aiming to minimize latency and energy consumption while maximizing throughput. Current efforts concentrate on optimizing architectures for specific model types, such as convolutional neural networks (CNNs) and transformers, often incorporating techniques like sparsity exploitation, quantization, and innovative memory structures (e.g., content-addressable memory). These advancements are crucial for deploying AI applications on resource-constrained edge devices and high-performance computing systems, impacting fields ranging from autonomous driving to natural language processing.
Papers
September 13, 2024
September 4, 2024
July 24, 2024
April 25, 2024
April 7, 2024
March 7, 2024
October 12, 2023
July 7, 2023
February 28, 2023
December 7, 2022
May 6, 2022
April 12, 2022