Fine Tuning
Fine-tuning adapts pre-trained large language models (LLMs) to specific tasks, improving performance and efficiency compared to training from scratch. Current research emphasizes efficient fine-tuning methods like low-rank adaptation (LoRA) and techniques addressing challenges such as catastrophic forgetting and calibration issues, often employing bilevel optimization or adaptive noise allocation for improved performance and privacy. This work is significant because it enables the deployment of powerful LLMs across diverse applications, from medical diagnosis to visual editing, while mitigating resource constraints and privacy concerns.
Papers
Offsite-Tuning: Transfer Learning without Full Model
Guangxuan Xiao, Ji Lin, Song Han
Knowledge is a Region in Weight Space for Fine-tuned Language Models
Almog Gueta, Elad Venezian, Colin Raffel, Noam Slonim, Yoav Katz, Leshem Choshen
Enhancing E-Commerce Recommendation using Pre-Trained Language Model and Fine-Tuning
Nuofan Xu, Chenhui Hu
PIRLNav: Pretraining with Imitation and RL Finetuning for ObjectNav
Ram Ramrakhya, Dhruv Batra, Erik Wijmans, Abhishek Das
Adapting Multilingual Speech Representation Model for a New, Underresourced Language through Multilingual Fine-tuning and Continued Pretraining
Karol Nowakowski, Michal Ptaszynski, Kyoko Murasaki, Jagna Nieuważny