LLaMA 3
LLaMA 3 is a family of large language models (LLMs) focusing on improving efficiency and performance across various tasks, including code generation, mathematical reasoning, and multilingual capabilities. Current research emphasizes techniques like parameter-efficient fine-tuning and selective layer enhancement to optimize knowledge injection and reduce computational costs, particularly for quantization and context length extension. These advancements are significant for expanding access to powerful LLMs and facilitating research in diverse scientific fields, as well as improving the efficiency and capabilities of practical applications.
Papers
October 27, 2024
October 16, 2024
October 11, 2024
October 3, 2024
September 27, 2024
September 25, 2024
September 10, 2024
August 27, 2024
June 12, 2024
June 11, 2024
May 6, 2024
May 1, 2024
April 30, 2024
April 22, 2024
April 5, 2024
January 4, 2024
December 26, 2023
December 14, 2023
September 12, 2023
August 24, 2023