Compact Model
Compact models aim to create smaller, faster, and more efficient machine learning models while maintaining or even exceeding the performance of larger counterparts. Current research focuses on techniques like knowledge distillation, weight multiplexing, and the integration of large language models to enhance smaller models' capabilities, particularly in low-data scenarios. These advancements are significant for deploying AI on resource-constrained devices and improving the efficiency of various applications, from natural language processing and image classification to device modeling and time-series analysis. The development of efficient training strategies tailored to compact architectures is also a key area of investigation.
Papers
October 7, 2024
October 1, 2024
August 22, 2024
July 22, 2024
June 24, 2024
May 11, 2024
April 17, 2024
March 22, 2024
February 2, 2024
November 10, 2023
August 3, 2023
June 1, 2023
May 2, 2023
October 18, 2022
October 4, 2022
June 30, 2022
May 19, 2022
April 14, 2022