Low Rank Decomposition
Low-rank decomposition is a technique for compressing large matrices and tensors by representing them as products of smaller matrices, thereby reducing computational cost and memory usage. Current research focuses on applying this to large neural networks, particularly in computer vision and natural language processing, using methods like singular value decomposition and Tucker decomposition, often integrated with pruning or other compression strategies. This work is driven by the need to deploy increasingly large models on resource-constrained devices and improve the efficiency of training and inference, impacting both the scalability of AI and its energy consumption.
Papers
October 17, 2024
October 15, 2024
September 20, 2024
August 29, 2024
August 12, 2024
July 24, 2024
July 5, 2024
June 28, 2024
May 28, 2024
May 24, 2024
May 10, 2024
May 6, 2024
April 30, 2024
March 28, 2024
March 10, 2024
December 12, 2023
December 10, 2023
December 1, 2023
October 17, 2023