Model Extraction Attack
Model extraction attacks aim to steal the functionality of machine learning models by querying their predictions, effectively replicating the model without access to its training data or internal parameters. Current research focuses on developing more efficient attack methods, particularly for large language models and object detectors, often employing techniques like knowledge distillation, active learning, and exploiting counterfactual explanations. This area is crucial for securing machine learning as a service platforms and protecting intellectual property, driving ongoing efforts to develop robust defenses such as watermarking and query unlearning.
Papers
September 4, 2024
July 1, 2024
June 14, 2024
May 29, 2024
May 8, 2024
April 28, 2024
April 4, 2024
March 15, 2024
March 5, 2024
March 3, 2024
January 26, 2024
December 22, 2023
December 17, 2023
December 8, 2023
November 23, 2023
November 10, 2023
October 25, 2023
October 21, 2023
October 15, 2023