Prompt Compression
Prompt compression aims to reduce the length of input prompts for large language models (LLMs) to improve computational efficiency and reduce costs without sacrificing performance. Research focuses on developing methods that selectively retain crucial information, employing techniques like extractive compression, summarization, and reinforcement learning to optimize compression ratios while preserving semantic meaning. These advancements are significant because they address the growing challenge of LLM resource consumption, enabling faster and more cost-effective deployment of these powerful models across various applications.
Papers
October 17, 2024
October 16, 2024
October 15, 2024
October 5, 2024
September 28, 2024
September 23, 2024
September 19, 2024
September 2, 2024
September 1, 2024
August 6, 2024
July 28, 2024
July 22, 2024
July 11, 2024
April 7, 2024
March 30, 2024
March 26, 2024
March 19, 2024
February 28, 2024
February 25, 2024