Multilingual Benchmark
Multilingual benchmarks are datasets designed to evaluate the performance of large language models (LLMs) across multiple languages, aiming to assess their cross-lingual capabilities and identify biases. Current research focuses on developing comprehensive benchmarks encompassing diverse tasks (e.g., question answering, code generation, translation) and languages, including low-resource ones, often employing instruction fine-tuning and various model architectures like transformers. These benchmarks are crucial for advancing the development of truly multilingual LLMs, improving their fairness and reliability, and enabling broader access to AI technologies across diverse linguistic communities.
Papers
EuroLLM: Multilingual Language Models for Europe
Pedro Henrique Martins, Patrick Fernandes, João Alves, Nuno M. Guerreiro, Ricardo Rei, Duarte M. Alves, José Pombal, Amin Farajian, Manuel Faysse, Mateusz Klimaszewski, Pierre Colombo, Barry Haddow, José G. C. de Souza, Alexandra Birch, André F. T. Martins
Unlocking Markets: A Multilingual Benchmark to Cross-Market Question Answering
Yifei Yuan, Yang Deng, Anders Søgaard, Mohammad Aliannejadi
Low-Resource Machine Translation through the Lens of Personalized Federated Learning
Viktor Moskvoretskii, Nazarii Tupitsa, Chris Biemann, Samuel Horváth, Eduard Gorbunov, Irina Nikishina
MultiSocial: Multilingual Benchmark of Machine-Generated Text Detection of Social-Media Texts
Dominik Macko, Jakub Kopal, Robert Moro, Ivan Srba