Linear Bandit
Linear bandits are a class of online learning problems where an agent sequentially selects actions (arms) from a set characterized by linear features, receiving stochastic rewards dependent on an unknown linear function of those features. Current research focuses on improving algorithm efficiency and robustness, exploring variations such as contextual bandits, incorporating human response times for preference learning, and addressing misspecified models or non-stationary environments. These advancements are significant for applications requiring efficient sequential decision-making under uncertainty, including personalized recommendations, clinical trials, and resource allocation, by enabling more accurate and adaptable algorithms.
Papers
May 26, 2022
May 22, 2022
May 12, 2022
May 3, 2022
April 3, 2022
March 29, 2022
March 8, 2022
March 2, 2022
February 28, 2022
February 23, 2022
February 21, 2022
February 14, 2022
February 12, 2022
February 7, 2022
February 4, 2022