Learned Embeddings
Learned embeddings are low-dimensional vector representations of data, aiming to capture semantic relationships and facilitate efficient processing for various tasks. Current research focuses on improving embedding quality through techniques like contrastive learning, attention mechanisms, and novel loss functions (e.g., centroid triplet loss), often within the context of specific model architectures such as transformers and graph neural networks. These advancements are driving progress in diverse fields, including computer vision (object identification, image generation), natural language processing (long-context language modeling, multi-hop reasoning), and recommendation systems, by enabling more accurate and efficient data analysis and downstream applications.
Papers
Smoothness and effective regularizations in learned embeddings for shape matching
Riccardo Marin, Souhaib Attaiki, Simone Melzi, Emanuele Rodolà, Maks Ovsjanikov
HET: Scaling out Huge Embedding Model Training via Cache-enabled Distributed Framework
Xupeng Miao, Hailin Zhang, Yining Shi, Xiaonan Nie, Zhi Yang, Yangyu Tao, Bin Cui