Adversarial Training Algorithm
Adversarial training aims to improve the robustness of deep neural networks against adversarial attacks, which involve subtly altering inputs to cause misclassification. Current research focuses on mitigating the trade-off between robustness and standard accuracy, exploring techniques like regularization (e.g., Fisher-Rao norm-based methods), refined min-max optimization strategies (e.g., focusing on "hiders"—previously defended samples that become vulnerable later), and incorporating unlabeled data. These advancements are crucial for deploying reliable machine learning models in safety-critical applications, where robustness to malicious inputs is paramount.
Papers
March 26, 2024
December 12, 2023
November 29, 2023
November 16, 2023
August 8, 2023
June 25, 2023
June 19, 2023
March 10, 2023
March 6, 2023
October 27, 2022
August 10, 2022
July 20, 2022
June 5, 2022
April 28, 2022
February 14, 2022
February 11, 2022