Privacy Attack

Privacy attacks target machine learning models to illicitly extract sensitive information from training data or model parameters. Current research focuses on developing and evaluating these attacks across various model architectures, including large language models, federated learning systems, and recommender systems, often employing techniques like membership inference, attribute inference, and model stealing. Understanding the effectiveness and limitations of these attacks is crucial for developing robust privacy-preserving techniques and ensuring responsible deployment of machine learning in sensitive applications. This research directly impacts the security and trustworthiness of AI systems, influencing both the design of privacy-preserving algorithms and the development of effective countermeasures.

Papers