Markov Decision Process
Markov Decision Processes (MDPs) are mathematical frameworks for modeling sequential decision-making problems under uncertainty, aiming to find optimal policies that maximize cumulative rewards. Current research emphasizes efficient algorithms for solving MDPs, particularly in complex settings like partially observable MDPs (POMDPs) and constrained MDPs (CMDPs), often employing techniques like policy gradient methods, Q-learning, and active inference. These advancements are crucial for improving the design and analysis of autonomous systems, robotics, and other applications requiring intelligent decision-making in dynamic environments, with a growing focus on addressing issues of safety, robustness, and sample efficiency.
Papers
October 16, 2024
October 15, 2024
October 14, 2024
October 10, 2024
October 9, 2024
October 8, 2024
October 7, 2024
October 3, 2024
October 2, 2024
September 27, 2024
September 26, 2024
September 25, 2024
September 23, 2024
September 22, 2024
September 21, 2024