Soft Actor Critic
Soft Actor-Critic (SAC) is a deep reinforcement learning algorithm aiming to learn robust and efficient policies by maximizing both expected reward and policy entropy. Current research focuses on improving SAC's sample efficiency, addressing safety constraints through methods like Lagrangian formulations and meta-gradient optimization, and extending its applicability to various domains including robotics, autonomous driving, and multi-agent systems. These advancements are significant because they enhance the practicality and reliability of reinforcement learning for real-world applications requiring safe and efficient decision-making in complex environments.
Papers
January 9, 2022
January 4, 2022
December 31, 2021
December 21, 2021
December 8, 2021
December 6, 2021