Many Sparse
Many Sparse research focuses on developing efficient methods for handling sparse data and models, primarily aiming to reduce computational costs and memory consumption while maintaining or improving performance. Current efforts concentrate on sparse neural network architectures (including Mixture-of-Experts models and various pruning techniques), sparse attention mechanisms in transformers, and sparse representations for various data types (e.g., point clouds, images). This work is significant for advancing machine learning applications in resource-constrained environments and enabling the scaling of large models to previously intractable sizes and complexities.
Papers
December 8, 2023
December 5, 2023
November 29, 2023
November 28, 2023
November 21, 2023
November 19, 2023
November 15, 2023
November 14, 2023
November 11, 2023
November 8, 2023
November 7, 2023
October 26, 2023
October 25, 2023
October 17, 2023
October 16, 2023
October 5, 2023
September 20, 2023
September 18, 2023