Many Sparse
Many Sparse research focuses on developing efficient methods for handling sparse data and models, primarily aiming to reduce computational costs and memory consumption while maintaining or improving performance. Current efforts concentrate on sparse neural network architectures (including Mixture-of-Experts models and various pruning techniques), sparse attention mechanisms in transformers, and sparse representations for various data types (e.g., point clouds, images). This work is significant for advancing machine learning applications in resource-constrained environments and enabling the scaling of large models to previously intractable sizes and complexities.
Papers
SPARF: Neural Radiance Fields from Sparse and Noisy Poses
Prune Truong, Marie-Julie Rakotosaona, Fabian Manhardt, Federico Tombari
Multiresolution kernel matrix algebra
H. Harbrecht, M. Multerer, O. Schenk, Ch. Schwab
SMAUG: Sparse Masked Autoencoder for Efficient Video-Language Pre-training
Yuanze Lin, Chen Wei, Huiyu Wang, Alan Yuille, Cihang Xie
FASTER-CE: Fast, Sparse, Transparent, and Robust Counterfactual Explanations
Shubham Sharma, Alan H. Gee, Jette Henderson, Joydeep Ghosh
The Lazy Neuron Phenomenon: On Emergence of Activation Sparsity in Transformers
Zonglin Li, Chong You, Srinadh Bhojanapalli, Daliang Li, Ankit Singh Rawat, Sashank J. Reddi, Ke Ye, Felix Chern, Felix Yu, Ruiqi Guo, Sanjiv Kumar