Linear Layer
Linear layers are fundamental components of neural networks, and research focuses on improving their efficiency and understanding their role in network behavior. Current efforts explore novel architectures like Block Tensor-Train Mixture-of-Experts (BTT-MoE) and Point Cloud Networks (PCN) to reduce computational costs and memory footprint, alongside techniques like low-rank decomposition and quantization to compress model size. These advancements are crucial for deploying large models on resource-constrained devices and improving the training and inference speed of large language models and other deep learning applications.
Papers
November 4, 2024
October 3, 2024
July 12, 2024
May 19, 2024
May 6, 2024
March 22, 2024
February 28, 2024
February 21, 2024
September 22, 2023
August 20, 2023
May 24, 2023
April 27, 2023
January 30, 2023
August 1, 2022
June 10, 2022
February 11, 2022
February 2, 2022
January 31, 2022