Mini Giant
"Mini-giant" research focuses on developing smaller, more efficient machine learning models that rival the performance of larger, computationally expensive counterparts. Current efforts concentrate on adapting existing architectures like transformers and convolutional neural networks, employing techniques such as model reprogramming and dynamic logit fusion to improve efficiency and generalization across diverse tasks, including image processing, natural language processing, and 3D modeling. This work is significant because it addresses the limitations of deploying large models in resource-constrained environments and promotes broader accessibility and reproducibility within the scientific community.
Papers
November 4, 2024
October 24, 2024
October 11, 2024
October 2, 2024
September 5, 2024
September 4, 2024
June 17, 2024
June 10, 2024
April 18, 2024
February 1, 2024
July 22, 2023
July 17, 2023
May 29, 2023
May 22, 2023
March 28, 2023