Attack Strategy
Attack strategies in machine learning and related fields focus on exploiting vulnerabilities in models and systems to achieve malicious objectives, such as data theft, model manipulation, or performance degradation. Current research emphasizes various attack types, including adversarial examples (e.g., crafted inputs causing misclassification), backdoor attacks (injecting triggers to control model outputs), and membership inference attacks (determining if a data point was used in training). These studies often involve deep neural networks, large language models, and reinforcement learning algorithms, and their findings are crucial for developing more robust and secure systems across diverse applications, from cybersecurity to AI safety.
Papers
December 12, 2023
December 8, 2023
November 23, 2023
November 8, 2023
October 24, 2023
September 29, 2023
August 11, 2023
June 29, 2023
June 19, 2023
June 15, 2023
June 8, 2023
May 31, 2023
May 24, 2023
May 18, 2023
April 21, 2023
March 19, 2023
February 16, 2023
February 1, 2023
January 23, 2023