Flatness Aware
Flatness-aware methods in machine learning aim to improve model generalization by finding flatter minima in the loss landscape, a characteristic empirically linked to better out-of-distribution performance. Current research focuses on understanding the relationship between flatness and transferability of adversarial examples, as well as the implicit biases introduced by optimizers like SGD in various learning paradigms (e.g., multitask, continual learning). These investigations, often employing deep linear networks or analyzing stochastic gradient descent dynamics, are crucial for developing more robust and generalizable machine learning models across diverse applications.
Papers
November 23, 2023
November 10, 2023
November 1, 2023
September 11, 2023
July 20, 2023
July 5, 2023
June 22, 2023
May 25, 2023
February 14, 2023