Jina Embeddings
Jina embeddings are vector representations of data, primarily text and images, designed to capture semantic meaning and relationships for improved information retrieval and downstream tasks. Current research focuses on enhancing embedding quality through novel loss functions (e.g., SimO loss for fine-grained contrastive learning), developing efficient architectures like decoupled embeddings for handling large datasets and multilingual contexts, and exploring non-Euclidean spaces (e.g., hyperbolic space) to better represent complex relationships. These advancements are improving performance in diverse applications, including recommendation systems, question answering, and even cybersecurity by enabling more accurate similarity searches and more effective model training.
Papers
On Debiasing Text Embeddings Through Context Injection
Thomas Uriot
Adversarially Robust Out-of-Distribution Detection Using Lyapunov-Stabilized Embeddings
Hossein Mirzaei, Mackenzie W. Mathis
Dissecting embedding method: learning higher-order structures from data
Liubov Tupikina (UPD5, LPI), Kathuria Hritika (LPI)
Detecting Unforeseen Data Properties with Diffusion Autoencoder Embeddings using Spine MRI data
Robert Graf, Florian Hunecke, Soeren Pohl, Matan Atad, Hendrik Moeller, Sophie Starck, Thomas Kroencke, Stefanie Bette, Fabian Bamberg, Tobias Pischon, Thoralf Niendorf, Carsten Schmidt, Johannes C. Paetzold, Daniel Rueckert, Jan S Kirschke
Gem: Gaussian Mixture Model Embeddings for Numerical Feature Distributions
Hafiz Tayyab Rauf, Alex Bogatu, Norman W. Paton, Andre Freitas
Siamese networks for Poincaré embeddings and the reconstruction of evolutionary trees
Ciro Carvallo, Hernán Bocaccio, Gabriel B. Mindlin, Pablo Groisman
Inference over Unseen Entities, Relations and Literals on Knowledge Graphs
Caglar Demir, N'Dah Jean Kouagou, Arnab Sharma, Axel-Cyrille Ngonga Ngomo
Stability of sorting based embeddings
Radu Balan, Efstratios Tsoukanis, Matthias Wellershoff
SimO Loss: Anchor-Free Contrastive Loss for Fine-Grained Supervised Contrastive Learning
Taha Bouhsine, Imad El Aaroussi, Atik Faysal, Wang Huaxia
DEPT: Decoupled Embeddings for Pre-training Language Models
Alex Iacob, Lorenzo Sani, Meghdad Kurmanji, William F. Shen, Xinchi Qiu, Dongqi Cai, Yan Gao, Nicholas D. Lane