LLM Fine Tuning
Fine-tuning large language models (LLMs) adapts pre-trained models to specific tasks using smaller datasets, improving performance and efficiency compared to training from scratch. Current research emphasizes parameter-efficient methods like LoRA and techniques to mitigate issues such as catastrophic forgetting and training data imbalance, often employing optimization algorithms like DPO and SVRG, and exploring diverse model architectures including Mixture-of-Experts. This area is crucial for deploying LLMs in real-world applications, enabling customization for various domains while addressing resource constraints and safety concerns.
Papers
February 18, 2024
February 9, 2024
February 8, 2024
February 5, 2024
February 4, 2024
January 8, 2024
January 1, 2024
December 29, 2023
December 19, 2023
December 17, 2023
November 12, 2023
October 30, 2023
October 6, 2023
September 29, 2023
September 22, 2023
September 16, 2023
September 2, 2023
September 1, 2023