Representation Learning
Representation learning aims to create meaningful and efficient data representations that capture underlying structure and facilitate downstream tasks like classification, prediction, and control. Current research focuses on developing robust and generalizable representations, often employing techniques like contrastive learning, transformers, and mixture-of-experts models, addressing challenges such as disentanglement, handling noisy or sparse data, and improving efficiency in multi-task and continual learning scenarios. These advancements have significant implications for various fields, improving the performance and interpretability of machine learning models across diverse applications, from recommendation systems to medical image analysis and causal inference.
Papers
Unsupervised Representation Learning in Partially Observable Atari Games
Li Meng, Morten Goodwin, Anis Yazidi, Paal Engelstad
The challenge of representation learning: Improved accuracy in deep vision models does not come with better predictions of perceptual similarity
Fritz Günther, Marco Marelli, Marco Alessandro Petilli
AMIGO: Sparse Multi-Modal Graph Transformer with Shared-Context Processing for Representation Learning of Giga-pixel Images
Ramin Nakhli, Puria Azadi Moghadam, Haoyang Mi, Hossein Farahani, Alexander Baras, Blake Gilks, Ali Bashashati
Can representation learning for multimodal image registration be improved by supervision of intermediate layers?
Elisabeth Wetzer, Joakim Lindblad, Nataša Sladoje