Input Token
Input tokens are the fundamental units of information processed by large language models (LLMs), particularly in multimodal applications combining text and visual data. Current research focuses on optimizing token representation and processing, including developing efficient encoding methods for visual information (e.g., using 2D features as tokens) and employing techniques like pruning and dynamic token selection to improve model efficiency without sacrificing performance. These advancements are crucial for deploying LLMs in resource-constrained environments and enhancing their capabilities in tasks like autonomous driving and automated captioning, where efficient processing of large amounts of data is paramount.
Papers
October 30, 2024
October 17, 2024
September 11, 2024
September 2, 2024
August 5, 2024
July 10, 2024
March 23, 2024
December 14, 2023
November 8, 2023
October 26, 2023