Smaller Language Model
Smaller language models (SLMs) aim to achieve comparable performance to their larger counterparts while significantly reducing computational costs and resource demands. Current research focuses on enhancing SLMs through techniques like knowledge distillation from larger models, dataset augmentation using LLMs, and innovative training methods such as contrastive fine-tuning and instruction tuning. These advancements are crucial for making advanced natural language processing capabilities accessible to a wider range of applications and researchers with limited resources, particularly in domains with constrained computational power or data availability.
Papers
May 24, 2023
May 23, 2023
May 17, 2023
April 14, 2023
December 20, 2022
December 19, 2022
December 1, 2022
October 8, 2022
September 2, 2022
July 10, 2022