Multilingual LLM
Multilingual Large Language Models (MLLMs) aim to create AI systems capable of understanding and generating text across multiple languages, overcoming the limitations of English-centric models. Current research focuses on improving performance in low-resource languages through techniques like chain-of-translation prompting, balanced multilingual datasets, and optimized multilingual tokenizers, often employing transformer-based architectures. These advancements are significant because they promote inclusivity in AI, enabling broader access to language technologies and facilitating cross-cultural communication and knowledge sharing.
Papers
Adapting Multilingual LLMs to Low-Resource Languages using Continued Pre-training and Synthetic Corpus
Raviraj Joshi, Kanishk Singla, Anusha Kamath, Raunak Kalani, Rakesh Paul, Utkarsh Vaidya, Sanjay Singh Chauhan, Niranjan Wartikar, Eileen Long
Towards Robust Knowledge Representations in Multilingual LLMs for Equivalence and Inheritance based Consistent Reasoning
Gaurav Arora, Srujana Merugu, Shreya Jain, Vaibhav Saxena