Defense Framework
Defense frameworks are being developed to protect machine learning models, particularly large language models and federated learning systems, from various adversarial attacks like data poisoning, model backdooring, and adversarial examples. Current research focuses on robust methods such as data curation, model merging, and homophily augmentation to improve model resilience, often employing techniques like mixture-of-experts models and outlier detection. These advancements are crucial for ensuring the reliability and trustworthiness of AI systems across diverse applications, ranging from natural language processing to autonomous driving and resource allocation, mitigating potential harms from malicious manipulations.
Papers
October 28, 2024
October 3, 2024
July 13, 2024
July 9, 2024
June 6, 2024
April 2, 2024
February 29, 2024
October 19, 2023
November 15, 2022
October 4, 2022
April 27, 2022