Decoder Only Large Language Model

Decoder-only large language models (LLMs) are a class of AI models designed for text generation, focusing on improving efficiency and leveraging pre-trained knowledge for various downstream tasks. Current research emphasizes efficient decoding methods, parameter-efficient fine-tuning techniques, and adapting these models for diverse applications like machine translation, speech-to-text translation, and knowledge graph construction, often involving techniques like prompt engineering and contrastive learning. This area is significant due to the potential for improved performance and reduced computational costs compared to encoder-decoder or encoder-only architectures, leading to broader accessibility and applicability in numerous fields.

Papers