Token Level
Token-level analysis in large language models (LLMs) focuses on understanding the individual units of text and their contribution to overall model behavior and performance. Current research investigates token dynamics within various architectures, including transformers and state space models, exploring techniques like token caching, selective training, and retrieval augmentation to improve efficiency and accuracy. This granular approach is crucial for enhancing LLM capabilities in diverse applications, from improving machine translation and gene expression prediction to mitigating biases and enhancing robustness against attacks. The insights gained are driving advancements in model training, optimization, and interpretability.
Papers
November 9, 2023
November 7, 2023
October 24, 2023
October 19, 2023
September 29, 2023
August 17, 2023
August 15, 2023
August 9, 2023
June 30, 2023
June 29, 2023
June 20, 2023
June 1, 2023
May 31, 2023
May 22, 2023
May 18, 2023
May 16, 2023
April 28, 2023
April 17, 2023
April 12, 2023