Contrastive Learning
Contrastive learning is a self-supervised machine learning technique that aims to learn robust data representations by contrasting similar and dissimilar data points. Current research focuses on applying contrastive learning to diverse modalities, including images, audio, text, and time-series data, often within multimodal frameworks and using architectures like MoCo and SimCLR, and exploring its application in various tasks such as object detection, speaker verification, and image dehazing. This approach is significant because it allows for effective learning from unlabeled or weakly labeled data, improving model generalization and performance across numerous applications, particularly in scenarios with limited annotated data or significant domain shifts.
Papers
Towards multi-modal anatomical landmark detection for ultrasound-guided brain tumor resection with contrastive learning
Soorena Salari, Amirhossein Rasoulian, Hassan Rivaz, Yiming Xiao
G2L: Semantically Aligned and Uniform Video Grounding via Geodesic and Game Theory
Hongxiang Li, Meng Cao, Xuxin Cheng, Yaowei Li, Zhihong Zhu, Yuexian Zou
Entropy Neural Estimation for Graph Contrastive Learning
Yixuan Ma, Xiaolin Zhang, Peng Zhang, Kun Zhan
Improving Semi-Supervised Semantic Segmentation with Dual-Level Siamese Structure Network
Zhibo Tain, Xiaolin Zhang, Peng Zhang, Kun Zhan
General-Purpose Multi-Modal OOD Detection Framework
Viet Duong, Qiong Wu, Zhengyi Zhou, Eric Zavesky, Jiahe Chen, Xiangzhou Liu, Wen-Ling Hsu, Huajie Shao
MARIO: Model Agnostic Recipe for Improving OOD Generalization of Graph Contrastive Learning
Yun Zhu, Haizhou Shi, Zhenshuo Zhang, Siliang Tang
Towards a Visual-Language Foundation Model for Computational Pathology
Ming Y. Lu, Bowen Chen, Drew F. K. Williamson, Richard J. Chen, Ivy Liang, Tong Ding, Guillaume Jaume, Igor Odintsov, Andrew Zhang, Long Phi Le, Georg Gerber, Anil V Parwani, Faisal Mahmood
CLIP-KD: An Empirical Study of CLIP Model Distillation
Chuanguang Yang, Zhulin An, Libo Huang, Junyu Bi, Xinqiang Yu, Han Yang, Boyu Diao, Yongjun Xu
PRIOR: Prototype Representation Joint Learning from Medical Images and Reports
Pujin Cheng, Li Lin, Junyan Lyu, Yijin Huang, Wenhan Luo, Xiaoying Tang
Homophily-Driven Sanitation View for Robust Graph Contrastive Learning
Yulin Zhu, Xing Ai, Yevgeniy Vorobeychik, Kai Zhou
Hallucination Improves the Performance of Unsupervised Visual Representation Learning
Jing Wu, Jennifer Hobbs, Naira Hovakimyan
Extracting Molecular Properties from Natural Language with Multimodal Contrastive Learning
Romain Lacombe, Andrew Gaut, Jeff He, David Lüdeke, Kateryna Pistunova
Contrastive Self-Supervised Learning Based Approach for Patient Similarity: A Case Study on Atrial Fibrillation Detection from PPG Signal
Subangkar Karmaker Shanto, Shoumik Saha, Atif Hasan Rahman, Mohammad Mehedy Masud, Mohammed Eunus Ali
Density-invariant Features for Distant Point Cloud Registration
Quan Liu, Hongzi Zhu, Yunsong Zhou, Hongyang Li, Shan Chang, Minyi Guo
Source-Free Domain Adaptation for Medical Image Segmentation via Prototype-Anchored Feature Alignment and Contrastive Learning
Qinji Yu, Nan Xi, Junsong Yuan, Ziyu Zhou, Kang Dang, Xiaowei Ding
Space Engage: Collaborative Space Supervision for Contrastive-based Semi-Supervised Semantic Segmentation
Changqi Wang, Haoyu Xie, Yuhui Yuan, Chong Fu, Xiangyu Yue