Sparse Mask
Sparse mask techniques aim to improve the efficiency and performance of neural networks by selectively activating or updating only a subset of parameters. Current research focuses on developing efficient algorithms for creating and applying these masks, particularly within transformer architectures and diffusion models, exploring various sparsity patterns (e.g., unstructured, structured, block-sparse) and incorporating them into training processes like preference optimization and federated learning. This work is significant because it addresses the computational cost and memory limitations of large models, leading to faster training, reduced inference times, and improved resource utilization in diverse applications.
Papers
October 7, 2024
September 23, 2024
September 13, 2024
August 25, 2024
August 20, 2024
June 25, 2024
June 4, 2024
April 16, 2024
April 9, 2024
December 11, 2023
September 19, 2023
June 30, 2023
May 27, 2023
April 6, 2023
February 13, 2023
February 2, 2023
December 27, 2022
October 11, 2022