Representation Learning
Representation learning aims to create meaningful and efficient data representations that capture underlying structure and facilitate downstream tasks like classification, prediction, and control. Current research focuses on developing robust and generalizable representations, often employing techniques like contrastive learning, transformers, and mixture-of-experts models, addressing challenges such as disentanglement, handling noisy or sparse data, and improving efficiency in multi-task and continual learning scenarios. These advancements have significant implications for various fields, improving the performance and interpretability of machine learning models across diverse applications, from recommendation systems to medical image analysis and causal inference.
Papers
Representation Learning via Variational Bayesian Networks
Oren Barkan, Avi Caciularu, Idan Rejwan, Ori Katz, Jonathan Weill, Itzik Malkiel, Noam Koenigstein
Semantic Positive Pairs for Enhancing Visual Representation Learning of Instance Discrimination methods
Mohammad Alkhalefi, Georgios Leontidis, Mingjun Zhong
Interpretable Anomaly Detection in Cellular Networks by Learning Concepts in Variational Autoencoders
Amandeep Singh, Michael Weber, Markus Lange-Hegermann
Hybrid Distillation: Connecting Masked Autoencoders with Contrastive Learners
Bowen Shi, Xiaopeng Zhang, Yaoming Wang, Jin Li, Wenrui Dai, Junni Zou, Hongkai Xiong, Qi Tian