Air Guardian System

"Air Guardian Systems" encompass a range of research efforts focused on enhancing the safety and reliability of large language models (LLMs) and other AI systems. Current research emphasizes developing robust guardrails against adversarial attacks (including jailbreaks), improving the accuracy and efficiency of hate speech detection, and ensuring responsible data usage in retrieval-augmented generation. These efforts utilize diverse approaches, including watermarking techniques, probabilistic graphical models for reasoning, and LLM-based agents for real-time monitoring and intervention, aiming to improve the trustworthiness and safety of AI in various applications. The ultimate goal is to mitigate risks associated with AI deployment, fostering greater user trust and responsible technological advancement.

Papers