Greedy Decoding
Greedy decoding is a fast, deterministic approach to generating text from large language models (LLMs), but its limitations in output quality have spurred research into improved methods. Current efforts focus on enhancing greedy decoding's performance through techniques like minimum Bayes risk decoding, which leverages external evaluators to select superior outputs, and self-ensembling methods that aggregate results from diverse prompts or multiple model runs. These advancements aim to improve the accuracy and efficiency of LLM text generation, impacting various applications from machine translation to speech recognition.
Papers
November 12, 2024
October 3, 2024
August 16, 2024
July 15, 2024
July 12, 2024
June 6, 2024
May 28, 2024
May 23, 2024
April 19, 2024
November 14, 2023
July 12, 2023