Simple Regret
Simple regret quantifies the difference between a learning algorithm's performance and the best possible outcome in hindsight, focusing on minimizing cumulative losses over a sequence of decisions. Current research investigates simple regret across diverse applications, including reinforcement learning, online optimization, and multi-agent systems, employing algorithms like Thompson Sampling, Follow-The-Perturbed-Leader, and various policy gradient methods, often within frameworks of online convex optimization. Understanding and minimizing simple regret is crucial for improving the efficiency and robustness of learning algorithms in dynamic environments, impacting fields ranging from resource allocation to personalized recommendations.
Papers
November 30, 2023
September 28, 2023
August 8, 2023
August 3, 2023
June 19, 2023
June 5, 2023
June 2, 2023
May 4, 2023
March 24, 2023
March 13, 2023
March 2, 2023
February 23, 2023
February 10, 2023
February 6, 2023
January 30, 2023
January 27, 2023
January 21, 2023
January 10, 2023
December 14, 2022