Contrastive Learning
Contrastive learning is a self-supervised machine learning technique that aims to learn robust data representations by contrasting similar and dissimilar data points. Current research focuses on applying contrastive learning to diverse modalities, including images, audio, text, and time-series data, often within multimodal frameworks and using architectures like MoCo and SimCLR, and exploring its application in various tasks such as object detection, speaker verification, and image dehazing. This approach is significant because it allows for effective learning from unlabeled or weakly labeled data, improving model generalization and performance across numerous applications, particularly in scenarios with limited annotated data or significant domain shifts.
Papers
On the Importance of Contrastive Loss in Multimodal Learning
Yunwei Ren, Yuanzhi Li
Anomalous Sound Detection using Audio Representation with Machine ID based Contrastive Learning Pretraining
Jian Guan, Feiyang Xiao, Youde Liu, Qiaoxi Zhu, Wenwu Wang
Linking Representations with Multimodal Contrastive Learning
Abhishek Arora, Xinmei Yang, Shao-Yu Jheng, Melissa Dell
Supervised Contrastive Learning with Heterogeneous Similarity for Distribution Shifts
Takuro Kutsuna
RegionPLC: Regional Point-Language Contrastive Learning for Open-World 3D Scene Understanding
Jihan Yang, Runyu Ding, Weipeng Deng, Zhe Wang, Xiaojuan Qi
MoLo: Motion-augmented Long-short Contrastive Learning for Few-shot Action Recognition
Xiang Wang, Shiwei Zhang, Zhiwu Qing, Changxin Gao, Yingya Zhang, Deli Zhao, Nong Sang
Focalized Contrastive View-invariant Learning for Self-supervised Skeleton-based Action Recognition
Qianhui Men, Edmond S. L. Ho, Hubert P. H. Shum, Howard Leung
Towards Integration of Discriminability and Robustness for Document-Level Relation Extraction
Jia Guo, Stanley Kok, Lidong Bing
Constructive Assimilation: Boosting Contrastive Learning Performance through View Generation Strategies
Ligong Han, Seungwook Han, Shivchander Sudalairaj, Charlotte Loh, Rumen Dangovski, Fei Deng, Pulkit Agrawal, Dimitris Metaxas, Leonid Karlinsky, Tsui-Wei Weng, Akash Srivastava
Learning with Fantasy: Semantic-Aware Virtual Contrastive Constraint for Few-Shot Class-Incremental Learning
Zeyin Song, Yifan Zhao, Yujun Shi, Peixi Peng, Li Yuan, Yonghong Tian
Towards Understanding the Mechanism of Contrastive Learning via Similarity Structure: A Theoretical Analysis
Hiroki Waida, Yuichiro Wada, Léo Andéol, Takumi Nakagawa, Yuhui Zhang, Takafumi Kanamori
HaLP: Hallucinating Latent Positives for Skeleton-based Self-Supervised Learning of Actions
Anshul Shah, Aniket Roy, Ketul Shah, Shlok Kumar Mishra, David Jacobs, Anoop Cherian, Rama Chellappa
Weakly-Supervised Text-driven Contrastive Learning for Facial Behavior Understanding
Xiang Zhang, Taoyue Wang, Xiaotian Li, Huiyuan Yang, Lijun Yin
SimTS: Rethinking Contrastive Representation Learning for Time Series Forecasting
Xiaochen Zheng, Xingyu Chen, Manuel Schürch, Amina Mollaysa, Ahmed Allam, Michael Krauthammer
LaCViT: A Label-aware Contrastive Fine-tuning Framework for Vision Transformers
Zijun Long, Zaiqiao Meng, Gerardo Aragon Camarasa, Richard McCreadie