Safety Critic
Safety critics are components of artificial intelligence systems, particularly in reinforcement learning, designed to evaluate the safety of an agent's actions and prevent unsafe behaviors. Current research focuses on developing effective safety critics using various methods, including gradient-based optimization, binary Bellman operators, and latent variable models within actor-critic frameworks. These advancements aim to improve the reliability and safety of AI agents in complex environments, addressing critical challenges in deploying AI in real-world applications such as robotics and autonomous systems.
Papers
June 23, 2024
May 24, 2024
January 23, 2024
June 24, 2023
October 2, 2022