Faster Training
Faster training of machine learning models is a crucial research area aiming to reduce computational costs and energy consumption while maintaining or improving model performance. Current efforts focus on optimizing existing architectures like Transformers and neural networks through techniques such as improved initialization strategies, efficient sampling methods (e.g., importance sampling, subgraph sampling), and architectural modifications (e.g., lightweight models, early exits, masked transformers). These advancements are significant because they enable the training of larger, more complex models and the application of deep learning to resource-constrained environments and time-sensitive tasks, ultimately accelerating progress across various scientific fields and practical applications.
Papers
SparseProp: Efficient Sparse Backpropagation for Faster Training of Neural Networks
Mahdi Nikdan, Tommaso Pegolotti, Eugenia Iofinova, Eldar Kurtic, Dan Alistarh
Liver Segmentation in Time-resolved C-arm CT Volumes Reconstructed from Dynamic Perfusion Scans using Time Separation Technique
Soumick Chatterjee, Hana Haseljić, Robert Frysch, Vojtěch Kulvait, Vladimir Semshchikov, Bennet Hensen, Frank Wacker, Inga Brüschx, Thomas Werncke, Oliver Speck, Andreas Nürnberger, Georg Rose