Self Supervised Learning
Self-supervised learning (SSL) aims to train machine learning models using unlabeled data by formulating pretext tasks that encourage the model to learn useful representations. Current research focuses on improving SSL's performance and generalization across diverse data types (images, audio, graphs, point clouds) and downstream tasks, employing techniques like contrastive learning, masked autoencoders, and generative models within various architectures such as transformers and convolutional neural networks. These advancements are significant because they reduce the reliance on expensive and time-consuming data labeling, enabling the development of robust models for applications ranging from medical image analysis and speech recognition to geospatial AI and protein function prediction. The efficiency gains from SSL are also a key focus, with research exploring optimal model and data sizes for given computational budgets.
Papers
Self-Supervised Learning for Building Robust Pediatric Chest X-ray Classification Models
Sheng Cheng, Zbigniew A. Starosolski, Devika Subramanian
Self-supervised learning for crystal property prediction via denoising
Alexander New, Nam Q. Le, Michael J. Pekala, Christopher D. Stiles
Contrastive Learning with Synthetic Positives
Dewen Zeng, Yawen Wu, Xinrong Hu, Xiaowei Xu, Yiyu Shi
Learning Valid Dual Bounds in Constraint Programming: Boosted Lagrangian Decomposition with Self-Supervised Learning
Swann Bessa, Darius Dabert, Max Bourgeat, Louis-Martin Rousseau, Quentin Cappart
Self-supervised Learning for Geospatial AI: A Survey
Yile Chen, Weiming Huang, Kaiqi Zhao, Yue Jiang, Gao Cong
Whitening Consistently Improves Self-Supervised Learning
András Kalapos, Bálint Gyires-Tóth
Protected Test-Time Adaptation via Online Entropy Matching: A Betting Approach
Yarin Bar, Shalev Shaer, Yaniv Romano
LiPCoT: Linear Predictive Coding based Tokenizer for Self-supervised Learning of Time Series Data via Language Models
Md Fahim Anjum
Efficient Test-Time Prompt Tuning for Vision-Language Models
Yuhan Zhu, Guozhen Zhang, Chen Xu, Haocheng Shen, Xiaoxin Chen, Gangshan Wu, Limin Wang
SMILES-Mamba: Chemical Mamba Foundation Models for Drug ADMET Prediction
Bohao Xu, Yingzhou Lu, Chenhao Li, Ling Yue, Xiao Wang, Nan Hao, Tianfan Fu, Jim Chen