Fine Tuning
Fine-tuning adapts pre-trained large language models (LLMs) to specific tasks, improving performance and efficiency compared to training from scratch. Current research emphasizes efficient fine-tuning methods like low-rank adaptation (LoRA) and techniques addressing challenges such as catastrophic forgetting and calibration issues, often employing bilevel optimization or adaptive noise allocation for improved performance and privacy. This work is significant because it enables the deployment of powerful LLMs across diverse applications, from medical diagnosis to visual editing, while mitigating resource constraints and privacy concerns.
Papers
Partial Fine-Tuning: A Successor to Full Fine-Tuning for Vision Transformers
Peng Ye, Yongqi Huang, Chongjun Tu, Minglei Li, Tao Chen, Tong He, Wanli Ouyang
A Split-and-Privatize Framework for Large Language Model Fine-Tuning
Xicong Shen, Yang Liu, Huiqi Liu, Jue Hong, Bing Duan, Zirui Huang, Yunlong Mao, Ye Wu, Di Wu
DRStageNet: Deep Learning for Diabetic Retinopathy Staging from Fundus Images
Yevgeniy Men, Jonathan Fhima, Leo Anthony Celi, Lucas Zago Ribeiro, Luis Filipe Nakayama, Joachim A. Behar
Hazards from Increasingly Accessible Fine-Tuning of Downloadable Foundation Models
Alan Chan, Ben Bucknall, Herbie Bradley, David Krueger
Carve3D: Improving Multi-view Reconstruction Consistency for Diffusion Models with RL Finetuning
Desai Xie, Jiahao Li, Hao Tan, Xin Sun, Zhixin Shu, Yi Zhou, Sai Bi, Sören Pirk, Arie E. Kaufman
Domain-Specific Fine-Tuning of Large Language Models for Interactive Robot Programming
Benjamin Alt, Urs Keßner, Aleksandar Taranovic, Darko Katic, Andreas Hermann, Rainer Jäkel, Gerhard Neumann
When Parameter-efficient Tuning Meets General-purpose Vision-language Models
Yihang Zhai, Haixin Wang, Jianlong Chang, Xinlong Yang, Jinan Sun, Shikun Zhang, Qi Tian
SPT: Fine-Tuning Transformer-based Language Models Efficiently with Sparsification
Yuntao Gui, Xiao Yan, Peiqi Yin, Han Yang, James Cheng