Efficient Training
Efficient training of large-scale machine learning models is a critical research area aiming to reduce computational costs and resource consumption while maintaining or improving model performance. Current efforts focus on optimizing training strategies for various architectures, including transformers, mixture-of-experts models, and neural operators, employing techniques like parameter-efficient fine-tuning, data pruning, and novel loss functions. These advancements are crucial for making advanced models like large language models and vision transformers more accessible and sustainable, impacting fields ranging from natural language processing and computer vision to scientific simulations and drug discovery.
Papers
April 17, 2023
April 7, 2023
March 27, 2023
March 23, 2023
March 21, 2023
February 22, 2023
February 8, 2023
February 2, 2023
January 30, 2023
January 23, 2023
December 23, 2022
December 8, 2022
November 25, 2022
November 17, 2022
October 19, 2022
October 13, 2022
September 5, 2022
August 4, 2022
July 28, 2022