Linear Bandit
Linear bandits are a class of online learning problems where an agent sequentially selects actions (arms) from a set characterized by linear features, receiving stochastic rewards dependent on an unknown linear function of those features. Current research focuses on improving algorithm efficiency and robustness, exploring variations such as contextual bandits, incorporating human response times for preference learning, and addressing misspecified models or non-stationary environments. These advancements are significant for applications requiring efficient sequential decision-making under uncertainty, including personalized recommendations, clinical trials, and resource allocation, by enabling more accurate and adaptable algorithms.
Papers
June 3, 2023
May 31, 2023
May 27, 2023
May 15, 2023
May 6, 2023
May 1, 2023
April 23, 2023
March 13, 2023
March 9, 2023
March 5, 2023
February 26, 2023
February 24, 2023
February 21, 2023
February 20, 2023
February 19, 2023
February 18, 2023
February 16, 2023
February 9, 2023