Token Level Contrastive
Token-level contrastive learning aims to improve model performance by enhancing the discriminability of individual word or sub-word representations (tokens). Current research focuses on applying this technique to various natural language processing and computer vision tasks, often incorporating it into existing architectures like BERT or SAM through novel algorithms that leverage contrastive losses at both token and sample levels. This approach shows promise in addressing challenges such as repetition in text generation, improving few-shot learning capabilities, and enhancing the alignment between different modalities (e.g., speech and text), leading to significant performance gains in diverse applications.
Papers
Tokenwise Contrastive Pretraining for Finer Speech-to-BERT Alignment in End-to-End Speech-to-Intent Systems
Vishal Sunder, Eric Fosler-Lussier, Samuel Thomas, Hong-Kwang J. Kuo, Brian Kingsbury
A Token-level Contrastive Framework for Sign Language Translation
Biao Fu, Peigen Ye, Liang Zhang, Pei Yu, Cong Hu, Yidong Chen, Xiaodong Shi