Decoder Only Large Language Model
Decoder-only large language models (LLMs) are a class of AI models designed for text generation, focusing on improving efficiency and leveraging pre-trained knowledge for various downstream tasks. Current research emphasizes efficient decoding methods, parameter-efficient fine-tuning techniques, and adapting these models for diverse applications like machine translation, speech-to-text translation, and knowledge graph construction, often involving techniques like prompt engineering and contrastive learning. This area is significant due to the potential for improved performance and reduced computational costs compared to encoder-decoder or encoder-only architectures, leading to broader accessibility and applicability in numerous fields.
Papers
February 7, 2024
January 25, 2024
January 21, 2024
December 30, 2023
November 6, 2023
October 17, 2023
September 29, 2023
July 5, 2023
May 23, 2023
April 13, 2023
February 2, 2023
January 29, 2023