Anomalous Token
Anomalous tokens, or tokens exhibiting unexpected behavior within large language models (LLMs), are a growing area of research focused on improving model reliability and trustworthiness. Current efforts concentrate on detecting these anomalies using techniques like low-confidence prediction analysis and clustering algorithms applied to attention patterns and embedding spaces, often within the context of specific model architectures such as Mixture-of-Experts (MoE). Identifying and mitigating the effects of anomalous tokens is crucial for enhancing the robustness and safety of LLMs, impacting both the development of more reliable AI systems and the broader understanding of their internal workings.
Papers
October 31, 2024
October 17, 2024
August 9, 2024
June 28, 2024
June 18, 2024
May 26, 2024
May 13, 2024
April 15, 2024
November 22, 2023
July 27, 2023
May 23, 2023
March 20, 2023
February 14, 2023