Linear Bandit
Linear bandits are a class of online learning problems where an agent sequentially selects actions (arms) from a set characterized by linear features, receiving stochastic rewards dependent on an unknown linear function of those features. Current research focuses on improving algorithm efficiency and robustness, exploring variations such as contextual bandits, incorporating human response times for preference learning, and addressing misspecified models or non-stationary environments. These advancements are significant for applications requiring efficient sequential decision-making under uncertainty, including personalized recommendations, clinical trials, and resource allocation, by enabling more accurate and adaptable algorithms.
Papers
May 12, 2024
May 9, 2024
May 5, 2024
March 15, 2024
March 10, 2024
March 1, 2024
February 24, 2024
February 20, 2024
February 19, 2024
February 12, 2024
February 11, 2024
January 21, 2024
January 17, 2024
January 14, 2024
November 14, 2023
November 8, 2023
October 23, 2023