Linear Bandit
Linear bandits are a class of online learning problems where an agent sequentially selects actions (arms) from a set characterized by linear features, receiving stochastic rewards dependent on an unknown linear function of those features. Current research focuses on improving algorithm efficiency and robustness, exploring variations such as contextual bandits, incorporating human response times for preference learning, and addressing misspecified models or non-stationary environments. These advancements are significant for applications requiring efficient sequential decision-making under uncertainty, including personalized recommendations, clinical trials, and resource allocation, by enabling more accurate and adaptable algorithms.
Papers
November 6, 2024
October 31, 2024
October 28, 2024
October 26, 2024
October 25, 2024
October 18, 2024
October 14, 2024
October 10, 2024
September 10, 2024
September 9, 2024
July 18, 2024
July 2, 2024
July 1, 2024
June 21, 2024
June 20, 2024
June 6, 2024
June 3, 2024
May 24, 2024
May 23, 2024