Sparsity Increase
Sparsity increase, a key focus in machine learning and signal processing, aims to reduce computational complexity and memory footprint by minimizing the number of non-zero elements in models or data representations. Current research explores various techniques, including algorithmic pruning, regularization methods (like L1 and Elastic Net), and novel architectures like sparse mixture-of-experts and sparsely activated neural networks, to achieve this goal across diverse applications. These advancements are significant because they enable efficient training and inference of large models, particularly on resource-constrained devices, and improve the interpretability and scalability of machine learning solutions in fields ranging from agriculture to medical diagnosis. The development of efficient algorithms for achieving and leveraging sparsity is a major driver of current research.
Papers
Pareto-optimal clustering with the primal deterministic information bottleneck
Andrew K. Tan, Max Tegmark, Isaac L. Chuang
SAFARI: Sparsity enabled Federated Learning with Limited and Unreliable Communications
Yuzhu Mao, Zihao Zhao, Meilin Yang, Le Liang, Yang Liu, Wenbo Ding, Tian Lan, Xiao-Ping Zhang