Supervised Finetuning
Supervised finetuning (SFT) adapts pre-trained large language models (LLMs) and other foundation models to specific tasks by training them on labeled data. Current research emphasizes improving SFT's efficiency, addressing issues like catastrophic forgetting (loss of pre-trained knowledge) and the high cost of data annotation, often exploring techniques like experimental design and parameter-efficient fine-tuning methods. This work is crucial for enhancing the performance and applicability of powerful models across diverse domains, from medical image analysis and speech recognition to improving the alignment of LLMs with human preferences and educational goals.
Papers
October 17, 2024
October 5, 2024
August 19, 2024
July 6, 2024
June 15, 2024
June 2, 2024
May 30, 2024
April 26, 2024
April 23, 2024
February 7, 2024
January 12, 2024
January 5, 2024
December 21, 2023
October 31, 2023
August 20, 2023
May 17, 2023
April 27, 2023
February 10, 2023
January 6, 2023