Weight Matrix
Weight matrices, the core components of neural networks, are the subject of intense research focused on improving efficiency, generalization, and interpretability. Current efforts explore low-rank approximations, structured matrices (e.g., Monarch, Block Tensor-Train), and novel training methods like weight decay and parameter-efficient fine-tuning (PEFT) techniques such as LoRA, to optimize their structure and reduce computational costs. These advancements are crucial for scaling up deep learning models, enabling their application to larger datasets and more complex tasks, and enhancing our understanding of how these models learn and generalize.
Papers
March 22, 2024
March 18, 2024
February 22, 2024
February 6, 2024
January 31, 2024
December 21, 2023
November 24, 2023
November 7, 2023
October 29, 2023
September 23, 2023
June 12, 2023
June 7, 2023
June 2, 2023
May 6, 2023
April 6, 2023
March 1, 2023
February 12, 2023
January 15, 2023
December 3, 2022