Token Level
Token-level analysis in large language models (LLMs) focuses on understanding the individual units of text and their contribution to overall model behavior and performance. Current research investigates token dynamics within various architectures, including transformers and state space models, exploring techniques like token caching, selective training, and retrieval augmentation to improve efficiency and accuracy. This granular approach is crucial for enhancing LLM capabilities in diverse applications, from improving machine translation and gene expression prediction to mitigating biases and enhancing robustness against attacks. The insights gained are driving advancements in model training, optimization, and interpretability.
Papers
Tokenwise Contrastive Pretraining for Finer Speech-to-BERT Alignment in End-to-End Speech-to-Intent Systems
Vishal Sunder, Eric Fosler-Lussier, Samuel Thomas, Hong-Kwang J. Kuo, Brian Kingsbury
Team \'UFAL at CMCL 2022 Shared Task: Figuring out the correct recipe for predicting Eye-Tracking features using Pretrained Language Models
Sunit Bhattacharya, Rishu Kumar, Ondrej Bojar
A Token-level Contrastive Framework for Sign Language Translation
Biao Fu, Peigen Ye, Liang Zhang, Pei Yu, Cong Hu, Yidong Chen, Xiaodong Shi