Self Supervised
Self-supervised learning (SSL) aims to train machine learning models using unlabeled data by designing pretext tasks that encourage the model to learn useful representations. Current research focuses on improving generalization, mitigating overfitting, and developing efficient architectures like transformers and CNNs for various modalities (images, audio, point clouds, fMRI data). SSL's significance lies in its ability to leverage vast amounts of readily available unlabeled data, leading to improved performance on downstream tasks and reducing the reliance on expensive and time-consuming manual labeling, particularly impacting fields like medical imaging, speech processing, and autonomous driving.
Papers
Exploring DINO: Emergent Properties and Limitations for Synthetic Aperture Radar Imagery
Joseph A. Gallego-Mejia, Anna Jungbluth, Laura Martínez-Ferrer, Matt Allen, Francisco Dorr, Freddie Kalaitzis, Raúl Ramos-Pollán
Machine learning the interaction network in coupled dynamical systems
Pawan R. Bhure, M. S. Santhanam
AstroCLIP: A Cross-Modal Foundation Model for Galaxies
Liam Parker, Francois Lanusse, Siavash Golkar, Leopoldo Sarra, Miles Cranmer, Alberto Bietti, Michael Eickenberg, Geraud Krawezik, Michael McCabe, Ruben Ohana, Mariel Pettee, Bruno Regaldo-Saint Blancard, Tiberiu Tesileanu, Kyunghyun Cho, Shirley Ho
FroSSL: Frobenius Norm Minimization for Efficient Multiview Self-Supervised Learning
Oscar Skean, Aayush Dhakal, Nathan Jacobs, Luis Gonzalo Sanchez Giraldo
XVO: Generalized Visual Odometry via Cross-Modal Self-Training
Lei Lai, Zhongkai Shangguan, Jimuyang Zhang, Eshed Ohn-Bar
CtxMIM: Context-Enhanced Masked Image Modeling for Remote Sensing Image Understanding
Mingming Zhang, Qingjie Liu, Yunhong Wang
Vision Transformers Need Registers
Timothée Darcet, Maxime Oquab, Julien Mairal, Piotr Bojanowski
Self-supervised Cross-view Representation Reconstruction for Change Captioning
Yunbin Tu, Liang Li, Li Su, Zheng-Jun Zha, Chenggang Yan, Qingming Huang
Joint Prediction and Denoising for Large-scale Multilingual Self-supervised Learning
William Chen, Jiatong Shi, Brian Yan, Dan Berrebbi, Wangyou Zhang, Yifan Peng, Xuankai Chang, Soumi Maiti, Shinji Watanabe
GasMono: Geometry-Aided Self-Supervised Monocular Depth Estimation for Indoor Scenes
Chaoqiang Zhao, Matteo Poggi, Fabio Tosi, Lei Zhou, Qiyu Sun, Yang Tang, Stefano Mattoccia
Elevating Skeleton-Based Action Recognition with Efficient Multi-Modality Self-Supervision
Yiping Wei, Kunyu Peng, Alina Roitberg, Jiaming Zhang, Junwei Zheng, Ruiping Liu, Yufan Chen, Kailun Yang, Rainer Stiefelhagen
Unlocking the Heart Using Adaptive Locked Agnostic Networks
Sylwia Majchrowska, Anders Hildeman, Philip Teare, Tom Diethe
DimCL: Dimensional Contrastive Learning For Improving Self-Supervised Learning
Thanh Nguyen, Trung Pham, Chaoning Zhang, Tung Luu, Thang Vu, Chang D. Yoo