Safe Agent
Safe agent research focuses on designing artificial intelligence agents that reliably and ethically perform tasks in real-world environments, mitigating risks such as adversarial attacks, bias, and unintended consequences. Current research emphasizes developing robust safety architectures, including input-output filters, safety agents, and hierarchical systems, often employing reinforcement learning algorithms like actor-critic methods and particle filters to achieve stability and optimize performance while adhering to safety constraints. This field is crucial for responsible AI deployment across various sectors, from autonomous vehicles to human-AI collaboration, ensuring both efficacy and safety in increasingly complex applications.
Papers
November 18, 2024
October 11, 2024
October 3, 2024
September 15, 2024
September 3, 2024
August 20, 2024
May 29, 2024
February 7, 2024
February 2, 2024
December 23, 2023
December 4, 2023
August 7, 2023
June 19, 2023
June 15, 2023
April 21, 2023
October 4, 2022
September 19, 2022
September 1, 2022