Multilingual Benchmark
Multilingual benchmarks are datasets designed to evaluate the performance of large language models (LLMs) across multiple languages, aiming to assess their cross-lingual capabilities and identify biases. Current research focuses on developing comprehensive benchmarks encompassing diverse tasks (e.g., question answering, code generation, translation) and languages, including low-resource ones, often employing instruction fine-tuning and various model architectures like transformers. These benchmarks are crucial for advancing the development of truly multilingual LLMs, improving their fairness and reliability, and enabling broader access to AI technologies across diverse linguistic communities.
Papers
Low-Resource Machine Translation through the Lens of Personalized Federated Learning
Viktor Moskvoretskii, Nazarii Tupitsa, Chris Biemann, Samuel Horváth, Eduard Gorbunov, Irina Nikishina
MultiSocial: Multilingual Benchmark of Machine-Generated Text Detection of Social-Media Texts
Dominik Macko, Jakub Kopal, Robert Moro, Ivan Srba