Fast LoRA
Fast LoRA (Low-Rank Adaptation) is a parameter-efficient fine-tuning technique for large language models (LLMs) that significantly reduces computational costs while maintaining performance comparable to full fine-tuning. Current research focuses on improving LoRA's efficiency and effectiveness through methods like tensor decomposition, selective aggregation (especially in federated learning settings), and novel optimization strategies to bridge the performance gap with full fine-tuning. These advancements are crucial for making LLMs more accessible and enabling their deployment on resource-constrained devices while addressing privacy concerns through federated learning approaches.
Papers
December 30, 2024
December 25, 2024
December 10, 2024
December 6, 2024
October 22, 2024
October 14, 2024
October 5, 2024
October 2, 2024
September 9, 2024
September 8, 2024
July 25, 2024
July 16, 2024
July 4, 2024
June 27, 2024
June 18, 2024
June 13, 2024
May 23, 2024
April 21, 2024