Supervised Fine Tuning
Supervised fine-tuning (SFT) adapts pre-trained large language models (LLMs) to specific tasks by training them on labeled data, aiming to improve performance and alignment with human preferences. Current research focuses on optimizing SFT methods, including exploring alternative loss functions (e.g., beyond cross-entropy), developing techniques to mitigate training imbalances and overfitting, and investigating the interplay between SFT and reinforcement learning. These advancements are significant because they enhance the efficiency and effectiveness of adapting LLMs for diverse applications, ranging from question answering and code generation to specialized domains like biomedicine and legal text processing.
Papers
September 20, 2023
September 1, 2023
August 23, 2023
August 14, 2023
August 3, 2023
July 26, 2023
July 20, 2023
July 5, 2023
June 14, 2023
May 23, 2023
May 9, 2023
April 4, 2023
November 28, 2022
November 14, 2022
September 30, 2022
September 13, 2022
August 11, 2022
June 1, 2022