Token Sequence
Token sequences, ordered sets of discrete units representing text or other data, are central to many large language model (LLM) applications. Current research focuses on optimizing token sequence handling for improved efficiency, security, and performance, exploring techniques like speculative decoding, variable-length training curricula, and novel tokenization methods (e.g., wavelet-based). These advancements aim to address challenges such as computational cost, vulnerability to adversarial attacks, and the efficient processing of long sequences, ultimately impacting the speed, accuracy, and safety of LLMs across diverse applications.
Papers
September 11, 2024
August 16, 2024
July 23, 2024
June 28, 2024
May 21, 2024
April 24, 2024
March 14, 2024
February 21, 2024
February 19, 2024
September 28, 2023
September 25, 2023
July 10, 2023
May 29, 2023
January 27, 2023
May 11, 2022