Multilingual Capability
Multilingual capability in large language models (LLMs) focuses on developing models that perform well across many languages, addressing the current dominance of English-centric systems. Research actively explores techniques like multilingual instruction tuning, continual pre-training, and manipulation of internal language representations to improve performance, particularly for low-resource languages, while mitigating issues like catastrophic forgetting and bias. This field is crucial for broadening AI accessibility globally and fostering equitable access to advanced AI services, impacting both scientific understanding of language representation and the development of inclusive real-world applications.
Papers
1+1>2: Can Large Language Models Serve as Cross-Lingual Knowledge Aggregators?
Yue Huang, Chenrui Fan, Yuan Li, Siyuan Wu, Tianyi Zhou, Xiangliang Zhang, Lichao Sun
Towards Truthful Multilingual Large Language Models: Benchmarking and Alignment Strategies
Weihao Liu, Ning Wu, Wenbiao Ding, Shining Liang, Ming Gong, Dongmei Zhang