Model Extraction Attack
Model extraction attacks aim to steal the functionality of machine learning models by querying their predictions, effectively replicating the model without access to its training data or internal parameters. Current research focuses on developing more efficient attack methods, particularly for large language models and object detectors, often employing techniques like knowledge distillation, active learning, and exploiting counterfactual explanations. This area is crucial for securing machine learning as a service platforms and protecting intellectual property, driving ongoing efforts to develop robust defenses such as watermarking and query unlearning.
Papers
October 3, 2023
August 31, 2023
August 9, 2023
July 3, 2023
June 8, 2023
May 17, 2023
April 25, 2023
April 17, 2023
April 6, 2023
March 25, 2023
March 13, 2023
February 16, 2023
February 6, 2023
February 4, 2023
November 29, 2022
November 24, 2022
September 14, 2022
July 27, 2022
July 19, 2022
June 28, 2022