Prompt Compression
Prompt compression aims to reduce the length of input prompts for large language models (LLMs) to improve computational efficiency and reduce costs without sacrificing performance. Research focuses on developing methods that selectively retain crucial information, employing techniques like extractive compression, summarization, and reinforcement learning to optimize compression ratios while preserving semantic meaning. These advancements are significant because they address the growing challenge of LLM resource consumption, enabling faster and more cost-effective deployment of these powerful models across various applications.
Papers
January 12, 2025
January 3, 2025
December 20, 2024
November 27, 2024
October 17, 2024
October 16, 2024
October 15, 2024
October 5, 2024
September 28, 2024
September 23, 2024
September 19, 2024
September 2, 2024
September 1, 2024
August 6, 2024
July 28, 2024
July 22, 2024
July 11, 2024
April 7, 2024
March 30, 2024