Fine Tuning
Fine-tuning adapts pre-trained large language models (LLMs) to specific tasks, improving performance and efficiency compared to training from scratch. Current research emphasizes efficient fine-tuning methods like low-rank adaptation (LoRA) and techniques addressing challenges such as catastrophic forgetting and calibration issues, often employing bilevel optimization or adaptive noise allocation for improved performance and privacy. This work is significant because it enables the deployment of powerful LLMs across diverse applications, from medical diagnosis to visual editing, while mitigating resource constraints and privacy concerns.
Papers
Initialization using Update Approximation is a Silver Bullet for Extremely Efficient Low-Rank Fine-Tuning
Kaustubh Ponkshe, Raghav Singhal, Eduard Gorbunov, Alexey Tumanov, Samuel Horvath, Praneeth Vepakomma
Quantized Delta Weight Is Safety Keeper
Yule Liu, Zhen Sun, Xinlei He, Xinyi Huang
Effective Fine-Tuning of Vision-Language Models for Accurate Galaxy Morphology Analysis
Ruoqi Wang, Haitao Wang, Qiong Luo
DreamBlend: Advancing Personalized Fine-tuning of Text-to-Image Diffusion Models
Shwetha Ram, Tal Neiman, Qianli Feng, Andrew Stuart, Son Tran, Trishul Chilimbi
Parameter-Efficient Transfer Learning for Music Foundation Models
Yiwei Ding, Alexander Lerch
PEFT-as-an-Attack! Jailbreaking Language Models during Federated Parameter-Efficient Fine-Tuning
Shenghui Li, Edith C.-H. Ngai, Fanghua Ye, Thiemo Voigt
Enhancing Parameter-Efficient Fine-Tuning of Vision Transformers through Frequency-Based Adaptation
Son Thai Ly, Hien V. Nguyen
Personalized Federated Fine-Tuning for LLMs via Data-Driven Heterogeneous Model Architectures
Yicheng Zhang, Zhen Qin, Zhaomin Wu, Shuiguang Deng
Evaluating Vision-Language Models as Evaluators in Path Planning
Mohamed Aghzal, Xiang Yue, Erion Plaku, Ziyu Yao
Fine-Tuning Large Language Models for Scientific Text Classification: A Comparative Study
Zhyar Rzgar K Rostam, Gábor Kertész
A gentle push funziona benissimo: making instructed models in Italian via contrastive activation steering
Daniel Scalena, Elisabetta Fersini, Malvina Nissim
Rapid Distributed Fine-tuning of a Segmentation Model Onboard Satellites
Meghan Plumridge, Rasmus Maråk, Chiara Ceccobello, Pablo Gómez, Gabriele Meoni, Filip Svoboda, Nicholas D. Lane
Not All Adapters Matter: Selective Adapter Freezing for Memory-Efficient Fine-Tuning of Language Models
Hyegang Son, Yonglak Son, Changhoon Kim, Young Geun Kim
Dynamic Self-Distillation via Previous Mini-batches for Fine-tuning Small Language Models
Yao Fu, Yin Yu, Xiaotian Han, Runchao Li, Xianxuan Long, Haotian Yu, Pan Li
DiffDesign: Controllable Diffusion with Meta Prior for Efficient Interior Design Generation
Yuxuan Yang, Jingyao Wang, Tao Geng, Wenwen Qiang, Changwen Zheng, Fuchun Sun
Parameter Efficient Instruction Tuning: An Empirical Study
Pengfei He