Fast LoRA
Fast LoRA (Low-Rank Adaptation) is a parameter-efficient fine-tuning technique for large language models (LLMs) that significantly reduces computational costs while maintaining performance comparable to full fine-tuning. Current research focuses on improving LoRA's efficiency and effectiveness through methods like tensor decomposition, selective aggregation (especially in federated learning settings), and novel optimization strategies to bridge the performance gap with full fine-tuning. These advancements are crucial for making LLMs more accessible and enabling their deployment on resource-constrained devices while addressing privacy concerns through federated learning approaches.
Papers
Improving LoRA in Privacy-preserving Federated Learning
Youbang Sun, Zitao Li, Yaliang Li, Bolin Ding
LoRA-Composer: Leveraging Low-Rank Adaptation for Multi-Concept Customization in Training-Free Diffusion Models
Yang Yang, Wen Wang, Liang Peng, Chaotian Song, Yao Chen, Hengjia Li, Xiaolong Yang, Qinglin Lu, Deng Cai, Boxi Wu, Wei Liu