Provable Benefit
Provable benefit research aims to rigorously establish the advantages of various machine learning techniques beyond empirical observation. Current efforts focus on theoretically analyzing the efficiency and effectiveness of methods like annealed Langevin Monte Carlo for sampling, sparse mixture of experts models with perturbed cosine routers, and curriculum learning for neural networks, often within specific problem settings (e.g., high-dimensional sparse Gaussian classification, reinforcement learning). These advancements provide a deeper understanding of algorithm performance, leading to improved model design and more reliable predictions in diverse applications, from reinforcement learning to quantum machine learning.
Papers
Gaussian Process Inference Using Mini-batch Stochastic Gradient Descent: Convergence Guarantees and Empirical Benefits
Hao Chen, Lili Zheng, Raed Al Kontar, Garvesh Raskutti
DyFormer: A Scalable Dynamic Graph Transformer with Provable Benefits on Generalization Ability
Weilin Cong, Yanhong Wu, Yuandong Tian, Mengting Gu, Yinglong Xia, Chun-cheng Jason Chen, Mehrdad Mahdavi