Multi Objective Reinforcement Learning
Multi-objective reinforcement learning (MORL) tackles the challenge of training agents to optimize multiple, often conflicting, objectives simultaneously. Current research focuses on developing efficient algorithms to discover the Pareto front—the set of optimal trade-offs between objectives—and improving the generalization of learned policies to unseen preferences, employing techniques like constrained optimization, preference inference from demonstrations, and novel architectures such as hypernets and diffusion models. These advancements are significant for addressing complex real-world problems in robotics, resource management, and other domains where single-objective approaches are insufficient, leading to more robust and adaptable AI systems.
Papers
In Search for Architectures and Loss Functions in Multi-Objective Reinforcement Learning
Mikhail Terekhov, Caglar Gulcehre
MOMAland: A Set of Benchmarks for Multi-Objective Multi-Agent Reinforcement Learning
Florian Felten, Umut Ucak, Hicham Azmani, Gao Peng, Willem Röpke, Hendrik Baier, Patrick Mannion, Diederik M. Roijers, Jordan K. Terry, El-Ghazali Talbi, Grégoire Danoy, Ann Nowé, Roxana Rădulescu