Many Sparse
Many Sparse research focuses on developing efficient methods for handling sparse data and models, primarily aiming to reduce computational costs and memory consumption while maintaining or improving performance. Current efforts concentrate on sparse neural network architectures (including Mixture-of-Experts models and various pruning techniques), sparse attention mechanisms in transformers, and sparse representations for various data types (e.g., point clouds, images). This work is significant for advancing machine learning applications in resource-constrained environments and enabling the scaling of large models to previously intractable sizes and complexities.
Papers
September 22, 2022
September 13, 2022
August 31, 2022
August 17, 2022
August 2, 2022
July 18, 2022
July 7, 2022
July 1, 2022
June 29, 2022
June 28, 2022
June 14, 2022
May 31, 2022
April 27, 2022
April 22, 2022
April 16, 2022
April 15, 2022
April 13, 2022
March 14, 2022