Information Geometry

Information geometry applies geometric concepts to probability distributions, aiming to understand and optimize machine learning and statistical inference processes. Current research focuses on leveraging information geometry's tools, such as Fisher information and Bregman divergences, to improve optimization algorithms (like Adam and natural gradient descent), develop novel generative models, and enhance the interpretability of neural networks by analyzing their parameter evolution during training. This framework offers a powerful mathematical lens for analyzing complex systems, leading to improved model efficiency, robustness, and a deeper understanding of learning dynamics across diverse fields.

Papers