Sentence Embeddings
Sentence embeddings represent sentences as dense vectors, aiming to capture their semantic meaning for various natural language processing tasks. Current research focuses on improving embedding quality through techniques like contrastive learning, domain adaptation (especially for low-resource languages), and exploring the internal structure of embeddings to better understand how linguistic information is encoded. These advancements are significant because effective sentence embeddings are crucial for applications ranging from semantic search and text classification to machine translation and recommendation systems.
Papers
DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings
Yung-Sung Chuang, Rumen Dangovski, Hongyin Luo, Yang Zhang, Shiyu Chang, Marin Soljačić, Shang-Wen Li, Wen-tau Yih, Yoon Kim, James Glass
SemEval-2022 Task 2: Multilingual Idiomaticity Detection and Sentence Embedding
Harish Tayyar Madabushi, Edward Gow-Smith, Marcos Garcia, Carolina Scarton, Marco Idiart, Aline Villavicencio
Is Neural Topic Modelling Better than Clustering? An Empirical Study on Clustering with Contextual Embeddings for Topics
Zihan Zhang, Meng Fang, Ling Chen, Mohammad-Reza Namazi-Rad