LLaMA Model
LLaMA (Large Language Model Meta AI) is a family of large language models known for their strong performance and open-source availability, driving research into efficient model architectures and safe deployment strategies. Current research focuses on optimizing LLaMA's efficiency through techniques like neural architecture search, quantization, and sparsity, aiming to reduce computational costs and improve inference speed without sacrificing accuracy. This work is significant because it facilitates broader access to powerful LLMs for various applications, while also addressing crucial safety concerns through the development of improved safety alignment datasets and techniques.
Papers
November 14, 2024
October 24, 2024
June 20, 2024
May 28, 2024
May 24, 2024
May 15, 2024
May 6, 2024
April 16, 2024
March 11, 2024
March 5, 2024
November 16, 2023
October 28, 2023
September 22, 2023
September 21, 2023
August 3, 2023
May 24, 2023
May 23, 2023
April 14, 2023
March 28, 2023