Linear Bandit
Linear bandits are a class of online learning problems where an agent sequentially selects actions (arms) from a set characterized by linear features, receiving stochastic rewards dependent on an unknown linear function of those features. Current research focuses on improving algorithm efficiency and robustness, exploring variations such as contextual bandits, incorporating human response times for preference learning, and addressing misspecified models or non-stationary environments. These advancements are significant for applications requiring efficient sequential decision-making under uncertainty, including personalized recommendations, clinical trials, and resource allocation, by enabling more accurate and adaptable algorithms.
Papers
October 9, 2023
October 4, 2023
October 3, 2023
October 1, 2023
September 25, 2023
September 1, 2023
August 29, 2023
July 27, 2023
July 24, 2023
July 5, 2023
June 26, 2023
June 15, 2023
June 13, 2023
June 12, 2023
June 3, 2023
May 31, 2023
May 27, 2023
May 15, 2023
May 6, 2023