Potential Scalability
Scalability in machine learning focuses on developing algorithms and architectures capable of efficiently handling massive datasets and complex models, addressing limitations of existing methods when dealing with increasingly large-scale data. Current research emphasizes techniques like distributed training for graph neural networks, efficient negative sampling strategies for extreme classification, and optimized algorithms for tasks such as recommendation systems and causal discovery, often employing novel architectures like Mamba and leveraging hardware acceleration (e.g., FPGAs and GPUs). These advancements are crucial for enabling the application of powerful machine learning models to real-world problems involving vast amounts of data, impacting fields ranging from scientific computing and personalized medicine to environmental monitoring and industrial automation.
Papers
Scalable Dynamic Mixture Model with Full Covariance for Probabilistic Traffic Forecasting
Seongjin Choi, Nicolas Saunier, Vincent Zhihao Zheng, Martin Trepanier, Lijun Sun
SMILE: Scaling Mixture-of-Experts with Efficient Bi-level Routing
Chaoyang He, Shuai Zheng, Aston Zhang, George Karypis, Trishul Chilimbi, Mahdi Soltanolkotabi, Salman Avestimehr
Aspects of scaling and scalability for flow-based sampling of lattice QCD
Ryan Abbott, Michael S. Albergo, Aleksandar Botev, Denis Boyda, Kyle Cranmer, Daniel C. Hackett, Alexander G. D. G. Matthews, Sébastien Racanière, Ali Razavi, Danilo J. Rezende, Fernando Romero-López, Phiala E. Shanahan, Julian M. Urban
An Analytics of Culture: Modeling Subjectivity, Scalability, Contextuality, and Temporality
Nanne van Noord, Melvin Wevers, Tobias Blanke, Julia Noordegraaf, Marcel Worring