Log Gradient
Log gradients, the gradients of the logarithm of a probability density function or likelihood, are a central concept in several machine learning subfields. Current research focuses on leveraging log gradients for improved algorithm efficiency and robustness, particularly in areas like reinforcement learning (via log density gradient methods), Bayesian neural networks (through Riemannian Laplace approximations), and causal discovery (using score matching techniques). These applications aim to address challenges such as sample complexity in reinforcement learning, non-Gaussian posteriors in Bayesian inference, and scalability in causal inference, ultimately leading to more efficient and reliable machine learning models.
Papers
August 26, 2024
May 27, 2024
March 3, 2024
September 5, 2023
June 12, 2023
April 6, 2023
February 21, 2023
July 13, 2022
April 20, 2022