Adversarial Direction
Adversarial direction research focuses on understanding and mitigating the vulnerability of deep learning models to adversarial attacks, which involve subtly manipulating inputs to cause misclassification. Current research explores methods to improve model robustness by leveraging geometric properties of data (e.g., tangent spaces), developing efficient algorithms for detecting and neutralizing adversarial perturbations in feature spaces, and employing strategies like adversarial unlearning to reduce overfitting. These advancements aim to enhance the reliability and security of deep learning systems across various applications, from image recognition to reinforcement learning.
Papers
October 22, 2024
August 27, 2024
August 23, 2024
June 9, 2023
April 26, 2023
April 25, 2023
April 4, 2023
December 30, 2022
December 9, 2022
October 28, 2022