Token Optimization
Token optimization focuses on reducing the number of tokens processed in large language models (LLMs) and vision transformers (ViTs) to improve efficiency and reduce computational costs. Current research explores various techniques, including adaptive token selection, joint optimization of tokens and model architecture (e.g., channel pruning), and the use of reinforcement learning algorithms like Proximal Policy Optimization (PPO) to learn optimal token-level strategies. These advancements are significant because they address limitations in current LLMs and ViTs, such as context window size and computational expense, leading to more efficient and cost-effective models for various applications.
Papers
Tokenphormer: Structure-aware Multi-token Graph Transformer for Node Classification
Zijie Zhou, Zhaoqi Lu, Xuekai Wei, Rongqin Chen, Shenghui Zhang, Pak Lon Ip, Leong Hou U
Token Preference Optimization with Self-Calibrated Visual-Anchored Rewards for Hallucination Mitigation
Jihao Gu, Yingyao Wang, Meng Cao, Pi Bu, Jun Song, Yancheng He, Shilong Li, Bo Zheng
Token Transformation Matters: Towards Faithful Post-hoc Explanation for Vision Transformer
Junyi Wu, Bin Duan, Weitai Kang, Hao Tang, Yan Yan
Open-Vocabulary Attention Maps with Token Optimization for Semantic Segmentation in Diffusion Models
Pablo Marcos-Manchón, Roberto Alcover-Couso, Juan C. SanMiguel, Jose M. Martínez