Information Theory
Information theory provides a mathematical framework for quantifying, storing, and communicating information, with applications spanning diverse fields. Current research focuses on leveraging information-theoretic principles—like mutual information and entropy—to improve machine learning algorithms, particularly in areas such as self-supervised learning, disentangled representation learning, and the design of robust and interpretable models (e.g., transformers). This involves developing new methods for estimating information-theoretic quantities in high-dimensional data and using these estimates to guide model design and analysis. The resulting advancements promise to enhance the efficiency, reliability, and explainability of AI systems across various applications, from healthcare diagnostics to material science.
Papers
Generalization Bounds: Perspectives from Information Theory and PAC-Bayes
Fredrik Hellström, Giuseppe Durisi, Benjamin Guedj, Maxim Raginsky
INSURE: An Information Theory Inspired Disentanglement and Purification Model for Domain Generalization
Xi Yu, Huan-Hsin Tseng, Shinjae Yoo, Haibin Ling, Yuewei Lin