Representation Learning
Representation learning aims to create meaningful and efficient data representations that capture underlying structure and facilitate downstream tasks like classification, prediction, and control. Current research focuses on developing robust and generalizable representations, often employing techniques like contrastive learning, transformers, and mixture-of-experts models, addressing challenges such as disentanglement, handling noisy or sparse data, and improving efficiency in multi-task and continual learning scenarios. These advancements have significant implications for various fields, improving the performance and interpretability of machine learning models across diverse applications, from recommendation systems to medical image analysis and causal inference.
Papers
Deep Impulse Responses: Estimating and Parameterizing Filters with Deep Networks
Alexander Richard, Peter Dodds, Vamsi Krishna Ithapu
Geometric Multimodal Contrastive Representation Learning
Petra Poklukar, Miguel Vasco, Hang Yin, Francisco S. Melo, Ana Paiva, Danica Kragic
Context Autoencoder for Self-Supervised Representation Learning
Xiaokang Chen, Mingyu Ding, Xiaodi Wang, Ying Xin, Shentong Mo, Yunhao Wang, Shumin Han, Ping Luo, Gang Zeng, Jingdong Wang
Discovering Distribution Shifts using Latent Space Representations
Leo Betthauser, Urszula Chajewska, Maurice Diesendruck, Rohith Pesala
Bootstrapped Representation Learning for Skeleton-Based Action Recognition
Olivier Moliner, Sangxia Huang, Kalle Åström
Learning Representation from Neural Fisher Kernel with Low-rank Approximation
Ruixiang Zhang, Shuangfei Zhai, Etai Littwin, Josh Susskind