Weight Matrix
Weight matrices, the core components of neural networks, are the subject of intense research focused on improving efficiency, generalization, and interpretability. Current efforts explore low-rank approximations, structured matrices (e.g., Monarch, Block Tensor-Train), and novel training methods like weight decay and parameter-efficient fine-tuning (PEFT) techniques such as LoRA, to optimize their structure and reduce computational costs. These advancements are crucial for scaling up deep learning models, enabling their application to larger datasets and more complex tasks, and enhancing our understanding of how these models learn and generalize.
Papers
June 22, 2022
June 12, 2022
April 5, 2022
April 1, 2022
March 28, 2022
March 8, 2022
March 3, 2022
February 11, 2022
December 31, 2021
November 26, 2021