Potential Vulnerability

Research into potential vulnerabilities focuses on identifying and mitigating weaknesses in various machine learning systems, including large language models (LLMs), federated learning frameworks, and AI agents. Current efforts concentrate on developing robust testing methodologies, such as reinforcement learning-based approaches and novel benchmark datasets, to uncover vulnerabilities like jailbreaks, malfunction amplification, and search space poisoning. Understanding these vulnerabilities is crucial for ensuring the security and reliability of increasingly prevalent AI systems across diverse applications, from smart grids to autonomous agents, and for building more trustworthy and resilient AI infrastructure.

Papers