Many Sparse
Many Sparse research focuses on developing efficient methods for handling sparse data and models, primarily aiming to reduce computational costs and memory consumption while maintaining or improving performance. Current efforts concentrate on sparse neural network architectures (including Mixture-of-Experts models and various pruning techniques), sparse attention mechanisms in transformers, and sparse representations for various data types (e.g., point clouds, images). This work is significant for advancing machine learning applications in resource-constrained environments and enabling the scaling of large models to previously intractable sizes and complexities.
Papers
May 28, 2024
May 24, 2024
May 23, 2024
May 17, 2024
May 15, 2024
May 8, 2024
May 2, 2024
April 27, 2024
April 16, 2024
April 11, 2024
April 5, 2024
March 29, 2024
March 26, 2024
March 21, 2024
March 14, 2024
March 13, 2024
March 7, 2024
February 29, 2024