Network Bottleneck
Network bottlenecks, limitations in data transfer speed or processing capacity within a system, hinder the performance of various applications, from large language models to federated learning and machine learning accelerators. Current research focuses on mitigating these bottlenecks through techniques like data compression (e.g., optimizing key-value cache usage in LLMs), algorithmic improvements (e.g., stochastic parameter updates in federated learning), and architectural changes (e.g., employing silicon photonic interposers in hardware accelerators). Overcoming these limitations is crucial for improving the efficiency, scalability, and performance of numerous computational systems and applications across diverse fields.
Papers
November 6, 2024
October 10, 2024
August 7, 2024
March 20, 2024
March 18, 2024
March 7, 2024
February 8, 2024
November 11, 2023
September 22, 2023