Equilibrium Bandit
Equilibrium bandit problems address the challenge of optimizing sequential decisions in systems that converge to an equilibrium state under a given action, where the system dynamics are unknown. Current research focuses on developing algorithms, such as Upper Equilibrium Concentration Bounds (UECB), that efficiently balance exploration of different actions with exploitation of near-equilibrium states, leveraging convergence bounds to minimize regret. This framework finds applications in diverse fields, including epidemic control, game theory, and molecular dynamics simulations, offering improved efficiency in learning optimal equilibria and enhancing the understanding of complex dynamic systems.
Papers
February 26, 2024
January 18, 2024
October 13, 2023
June 8, 2023
February 27, 2023
February 15, 2023
October 24, 2022
June 30, 2022