Loss Gradient
Loss gradients, the rate of change of a loss function with respect to model parameters, are central to training machine learning models. Current research focuses on mitigating challenges arising from loss gradients, such as conflicting gradients in multi-task learning and imbalanced gradients in long-tailed datasets, often employing techniques like gradient surgery, dynamic weighting, and alternate training procedures across various architectures including convolutional neural networks and transformers. These efforts aim to improve model training efficiency, generalization, and robustness, impacting diverse applications from image segmentation and object detection to physics-informed neural networks and energy-efficient deep learning.
Papers
October 20, 2024
August 20, 2024
July 1, 2024
June 11, 2024
May 7, 2024
April 10, 2024
February 27, 2024
February 5, 2024
February 3, 2024
December 26, 2023
August 31, 2023
August 29, 2023
August 19, 2023
October 11, 2022
August 7, 2022
July 15, 2022
May 19, 2022
March 21, 2022