Better Adversarial Robustness
Improving the adversarial robustness of machine learning models is a crucial research area focused on making these models less susceptible to malicious attacks that manipulate inputs to cause misclassification. Current efforts explore diverse approaches, including novel model architectures like diffusion classifiers and part-based models, refined adversarial training techniques (e.g., trajectory reweighting and curvature regularization), and innovative defense strategies such as dataset distillation and combined transduction/rejection methods. These advancements are vital for enhancing the reliability and security of machine learning systems across various applications, from image recognition to natural language processing.
Papers
August 16, 2024
July 15, 2024
March 15, 2024
June 25, 2023
May 27, 2023
May 25, 2023
December 5, 2022
November 8, 2022
August 17, 2022