Average Approximation
Average approximation focuses on efficiently representing complex functions or data using simpler models, aiming to balance accuracy with computational tractability. Current research explores this through various approaches, including low-rank matrix approximations for efficient parameter estimation in large language models and neural networks, and the development of novel algorithms like adaptive proximal gradient methods for optimization under relaxed smoothness assumptions. These advancements have significant implications for diverse fields, improving the efficiency and scalability of machine learning algorithms, enhancing the interpretability of complex models, and enabling real-time applications in areas like robotics and control systems.
Papers
Mnemosyne: Parallelization Strategies for Efficiently Serving Multi-Million Context Length LLM Inference Requests Without Approximations
Amey Agrawal, Junda Chen, Íñigo Goiri, Ramachandran Ramjee, Chaojie Zhang, Alexey Tumanov, Esha Choukse
Monge-Kantorovich Fitting With Sobolev Budgets
Forest Kobayashi, Jonathan Hayase, Young-Heon Kim