Large Pre Trained Model
Large pre-trained models (LPMs) are massive neural networks trained on enormous datasets, aiming to achieve strong generalization across diverse downstream tasks with minimal further training. Current research emphasizes efficient fine-tuning techniques, such as prompt engineering, low-rank adaptation (e.g., LoRA, SVFit), and sparse parameter updates, to reduce computational costs and improve model adaptability while addressing issues like overfitting and catastrophic forgetting. This field is significant due to LPMs' transformative impact on various applications, from natural language processing and computer vision to robotics and education, driving advancements in both theoretical understanding and practical deployment of AI systems.
Papers
May 26, 2023
May 16, 2023
May 8, 2023
April 29, 2023
April 19, 2023
April 13, 2023
March 19, 2023
March 16, 2023
March 13, 2023
March 2, 2023
January 22, 2023
December 23, 2022
December 21, 2022
November 28, 2022
November 3, 2022
October 26, 2022
October 20, 2022
October 17, 2022
October 13, 2022