Sparse Format
Sparse formats are data structures designed to efficiently represent matrices and tensors with a high proportion of zero values, a common characteristic in many large-scale machine learning models. Current research focuses on developing novel sparse formats optimized for specific applications, such as large language model inference and spiking neural network simulation, often incorporating techniques like compression and tailored code generation for improved performance on diverse hardware architectures. These advancements aim to reduce memory consumption and accelerate computations, thereby enabling the training and deployment of larger and more complex models while addressing the limitations of existing general-purpose sparse formats.
Papers
July 23, 2024
June 17, 2024
March 9, 2024
November 28, 2023
September 8, 2023
September 6, 2023
April 12, 2023