Fine Tuning
Fine-tuning adapts pre-trained large language models (LLMs) to specific tasks, improving performance and efficiency compared to training from scratch. Current research emphasizes efficient fine-tuning methods like low-rank adaptation (LoRA) and techniques addressing challenges such as catastrophic forgetting and calibration issues, often employing bilevel optimization or adaptive noise allocation for improved performance and privacy. This work is significant because it enables the deployment of powerful LLMs across diverse applications, from medical diagnosis to visual editing, while mitigating resource constraints and privacy concerns.
Papers
DEFT: Efficient Finetuning of Conditional Diffusion Models by Learning the Generalised $h$-transform
Alexander Denker, Francisco Vargas, Shreyas Padhy, Kieran Didi, Simon Mathis, Vincent Dutordoir, Riccardo Barbano, Emile Mathieu, Urszula Julia Komorowska, Pietro Lio
LoFiT: Localized Fine-tuning on LLM Representations
Fangcong Yin, Xi Ye, Greg Durrett
Automatic Essay Multi-dimensional Scoring with Fine-tuning and Multiple Regression
Kun Sun, Rong Wang
Empirical influence functions to understand the logic of fine-tuning
Jordan K. Matelsky, Lyle Ungar, Konrad P. Kording
The Best of Both Worlds: Toward an Honest and Helpful Large Language Model
Chujie Gao, Qihui Zhang, Dongping Chen, Yue Huang, Siyuan Wu, Zhengyan Fu, Yao Wan, Xiangliang Zhang, Lichao Sun
Mamba State-Space Models Are Lyapunov-Stable Learners
John T. Halloran, Manbir Gulati, Paul F. Roysdon
QuanTA: Efficient High-Rank Fine-Tuning of LLMs with Quantum-Informed Tensor Adaptation
Zhuo Chen, Rumen Dangovski, Charlotte Loh, Owen Dugan, Di Luo, Marin Soljačić
Enhancing Generative Molecular Design via Uncertainty-guided Fine-tuning of Variational Autoencoders
A N M Nafiz Abeer, Sanket Jantre, Nathan M Urban, Byung-Jun Yoon
Transfer Q Star: Principled Decoding for LLM Alignment
Souradip Chakraborty, Soumya Suvra Ghosal, Ming Yin, Dinesh Manocha, Mengdi Wang, Amrit Singh Bedi, Furong Huang
ETHER: Efficient Finetuning of Large-Scale Models with Hyperplane Reflections
Massimo Bini, Karsten Roth, Zeynep Akata, Anna Khoreva
TAIA: Large Language Models are Out-of-Distribution Data Learners
Shuyang Jiang, Yusheng Liao, Ya Zhang, Yanfeng Wang, Yu Wang
The Fine-Tuning Paradox: Boosting Translation Quality Without Sacrificing LLM Abilities
David Stap, Eva Hasler, Bill Byrne, Christof Monz, Ke Tran
Exploring Diffusion Models' Corruption Stage in Few-Shot Fine-tuning and Mitigating with Bayesian Neural Networks
Xiaoyu Wu, Jiaru Zhang, Yang Hua, Bohan Lyu, Hao Wang, Tao Song, Haibing Guan
Ensemble Model With Bert,Roberta and Xlnet For Molecular property prediction
Junling Hu
STAT: Shrinking Transformers After Training
Megan Flynn, Alexander Wang, Dean Edward Alvarez, Christopher De Sa, Anil Damle
Contrastive-Adversarial and Diffusion: Exploring pre-training and fine-tuning strategies for sulcal identification
Michail Mamalakis, Héloïse de Vareilles, Shun-Chin Jim Wu, Ingrid Agartz, Lynn Egeland Mørch-Johnsen, Jane Garrison, Jon Simons, Pietro Lio, John Suckling, Graham Murray
DGRC: An Effective Fine-tuning Framework for Distractor Generation in Chinese Multi-choice Reading Comprehension
Runfeng Lin, Dacheng Xu, Huijiang Wang, Zebiao Chen, Yating Wang, Shouqiang Liu