Large Deep Learning Model
Large deep learning models, characterized by their massive size and impressive performance across diverse tasks, are a central focus of current research. Efforts concentrate on improving training efficiency through techniques like model parallelism, quantization (including post-training and quantization-aware training), and novel optimization algorithms, while simultaneously addressing memory constraints and reducing computational costs. This research is crucial for making these powerful models more accessible and sustainable, impacting various fields from natural language processing and computer vision to remote sensing and scientific computing.
Papers
October 19, 2024
October 14, 2024
September 18, 2024
August 3, 2024
June 11, 2024
April 17, 2024
April 16, 2024
March 13, 2024
February 23, 2024
January 25, 2024
January 19, 2024
January 3, 2024
October 31, 2023
September 29, 2023
September 3, 2023
April 27, 2023
April 19, 2023
April 17, 2023
March 25, 2023