Cross Lingual
Cross-lingual research focuses on bridging language barriers in natural language processing, aiming to build models that understand and process text across multiple languages. Current efforts concentrate on improving multilingual large language models (LLMs) through techniques like continual pre-training, adapter modules, and contrastive learning, often addressing challenges related to low-resource languages and semantic alignment. This field is crucial for expanding access to NLP technologies globally and enabling cross-cultural communication and information exchange in diverse applications, such as machine translation, sentiment analysis, and cross-lingual information retrieval.
Papers
Self-Distillation for Model Stacking Unlocks Cross-Lingual NLU in 200+ Languages
Fabian David Schmidt, Philipp Borchert, Ivan Vulić, Goran Glavaš
MultiSocial: Multilingual Benchmark of Machine-Generated Text Detection of Social-Media Texts
Dominik Macko, Jakub Kopal, Robert Moro, Ivan Srba
Exploring Intra and Inter-language Consistency in Embeddings with ICA
Rongzhi Li, Takeru Matsuda, Hitomi Yanaka