Self Supervised Learning
Self-supervised learning (SSL) aims to train machine learning models using unlabeled data by formulating pretext tasks that encourage the model to learn useful representations. Current research focuses on improving SSL's performance and generalization across diverse data types (images, audio, graphs, point clouds) and downstream tasks, employing techniques like contrastive learning, masked autoencoders, and generative models within various architectures such as transformers and convolutional neural networks. These advancements are significant because they reduce the reliance on expensive and time-consuming data labeling, enabling the development of robust models for applications ranging from medical image analysis and speech recognition to geospatial AI and protein function prediction. The efficiency gains from SSL are also a key focus, with research exploring optimal model and data sizes for given computational budgets.
Papers
Stroke-Based Autoencoders: Self-Supervised Learners for Efficient Zero-Shot Chinese Character Recognition
Zongze Chen, Wenxia Yang, Xin Li
Federated Self-Supervised Learning in Heterogeneous Settings: Limits of a Baseline Approach on HAR
Sannara Ek, Romain Rombourg, François Portet, Philippe Lalanda
Self-Supervised-RCNN for Medical Image Segmentation with Limited Data Annotation
Banafshe Felfeliyan, Abhilash Hareendranathan, Gregor Kuntze, David Cornell, Nils D. Forkert, Jacob L. Jaremko, Janet L. Ronsky
Multi-Modal Unsupervised Pre-Training for Surgical Operating Room Workflow Analysis
Muhammad Abdullah Jamal, Omid Mohareri
On the Importance of Hyperparameters and Data Augmentation for Self-Supervised Learning
Diane Wagner, Fabio Ferreira, Danny Stoll, Robin Tibor Schirrmeister, Samuel Müller, Frank Hutter
Label-Efficient Self-Supervised Speaker Verification With Information Maximization and Contrastive Learning
Théo Lepage, Réda Dehak
Synergistic Self-supervised and Quantization Learning
Yun-Hao Cao, Peiqin Sun, Yechang Huang, Jianxin Wu, Shuchang Zhou
Dual Contrastive Learning for Spatio-temporal Representation
Shuangrui Ding, Rui Qian, Hongkai Xiong