Better Adversarial Robustness

Improving the adversarial robustness of machine learning models is a crucial research area focused on making these models less susceptible to malicious attacks that manipulate inputs to cause misclassification. Current efforts explore diverse approaches, including novel model architectures like diffusion classifiers and part-based models, refined adversarial training techniques (e.g., trajectory reweighting and curvature regularization), and innovative defense strategies such as dataset distillation and combined transduction/rejection methods. These advancements are vital for enhancing the reliability and security of machine learning systems across various applications, from image recognition to natural language processing.

Papers