Potential Vulnerability
Research into potential vulnerabilities focuses on identifying and mitigating weaknesses in various machine learning systems, including large language models (LLMs), federated learning frameworks, and AI agents. Current efforts concentrate on developing robust testing methodologies, such as reinforcement learning-based approaches and novel benchmark datasets, to uncover vulnerabilities like jailbreaks, malfunction amplification, and search space poisoning. Understanding these vulnerabilities is crucial for ensuring the security and reliability of increasingly prevalent AI systems across diverse applications, from smart grids to autonomous agents, and for building more trustworthy and resilient AI infrastructure.
Papers
September 16, 2024
July 30, 2024
July 11, 2024
June 26, 2024
June 12, 2024
March 19, 2024
December 13, 2023
December 16, 2022
September 20, 2022
July 5, 2022
June 12, 2022