Instruction Tuned Large Language Model
Instruction-tuned large language models (LLMs) are designed to improve the ability of LLMs to follow instructions accurately and generate relevant responses, addressing limitations of previous models. Current research focuses on improving instruction-following capabilities through techniques like continual pretraining, model merging, and reinforcement learning from human feedback, often applied to architectures such as Llama and GPT models. This area is significant because it enhances the reliability and safety of LLMs for various applications, including finance, healthcare, and software development, while also raising important questions about bias mitigation and robustness to adversarial attacks.
Papers
September 25, 2023
September 19, 2023
September 7, 2023
September 4, 2023
August 28, 2023
August 27, 2023
August 17, 2023
August 6, 2023
July 31, 2023
July 29, 2023
June 19, 2023
June 5, 2023
May 23, 2023
May 18, 2023
April 27, 2023
April 24, 2023
April 20, 2023
February 28, 2023
February 11, 2023