Weight Matrix
Weight matrices, the core components of neural networks, are the subject of intense research focused on improving efficiency, generalization, and interpretability. Current efforts explore low-rank approximations, structured matrices (e.g., Monarch, Block Tensor-Train), and novel training methods like weight decay and parameter-efficient fine-tuning (PEFT) techniques such as LoRA, to optimize their structure and reduce computational costs. These advancements are crucial for scaling up deep learning models, enabling their application to larger datasets and more complex tasks, and enhancing our understanding of how these models learn and generalize.
Papers
November 1, 2024
October 5, 2024
October 3, 2024
October 2, 2024
September 23, 2024
August 5, 2024
July 16, 2024
June 10, 2024
June 7, 2024
May 11, 2024
April 4, 2024
March 22, 2024
March 18, 2024
February 22, 2024
February 6, 2024
January 31, 2024
December 21, 2023
November 24, 2023
November 7, 2023