Text Embeddings
Text embeddings are numerical representations of text that capture semantic meaning, enabling computers to understand and process language. Current research focuses on improving the quality and controllability of these embeddings, particularly through techniques like contrastive learning, fine-tuning large language models (LLMs), and developing novel architectures to better handle complex prompts and disentangle attributes within embeddings. These advancements are crucial for various applications, including image generation, information retrieval, and sentiment analysis, improving the performance and efficiency of numerous natural language processing tasks.
Papers
A Large-Scale Sensitivity Analysis on Latent Embeddings and Dimensionality Reductions for Text Spatializations
Daniel Atzberger, Tim Cech, Willy Scheibel, Jürgen Döllner, Michael Behrisch, Tobias Schreck
DragText: Rethinking Text Embedding in Point-based Image Editing
Gayoon Choi, Taejin Jeong, Sujung Hong, Seong Jae Hwang
Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe
Alicja Ziarko, Albert Q. Jiang, Bartosz Piotrowski, Wenda Li, Mateja Jamnik, Piotr Miłoś
LLMEmbed: Rethinking Lightweight LLM's Genuine Function in Text Classification
Chun Liu, Hongguang Zhang, Kainan Zhao, Xinghai Ju, Lin Yang