Model Stealing
Model stealing involves adversaries illicitly replicating machine learning models by exploiting query access or leaked information, thereby undermining intellectual property and potentially compromising sensitive data. Current research focuses on developing robust defenses against these attacks, particularly for large language models and self-supervised learning models, employing techniques like hardware-based restrictions, obfuscation of model architectures, and watermarking. This active area of research is crucial for securing the deployment of machine learning models as a service and protecting the valuable intellectual property embedded within them.
Papers
November 12, 2024
May 31, 2024
April 18, 2024
April 17, 2024
February 26, 2024
November 8, 2023
October 17, 2023
September 29, 2023
September 18, 2023
May 9, 2023
April 13, 2023
September 16, 2022
August 4, 2022
June 16, 2022
April 23, 2022
February 21, 2022
January 31, 2022
January 27, 2022
January 23, 2022